var/home/core/zuul-output/0000755000175000017500000000000015067172450014534 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015067203173015475 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004761447315067203165017722 0ustar rootrootOct 01 09:36:06 crc systemd[1]: Starting Kubernetes Kubelet... Oct 01 09:36:06 crc restorecon[4664]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 01 09:36:06 crc restorecon[4664]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 09:36:07 crc restorecon[4664]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 01 09:36:07 crc restorecon[4664]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 01 09:36:08 crc kubenswrapper[4787]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 01 09:36:08 crc kubenswrapper[4787]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 01 09:36:08 crc kubenswrapper[4787]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 01 09:36:08 crc kubenswrapper[4787]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 01 09:36:08 crc kubenswrapper[4787]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 01 09:36:08 crc kubenswrapper[4787]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.266613 4787 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272568 4787 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272606 4787 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272614 4787 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272620 4787 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272625 4787 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272631 4787 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272637 4787 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272643 4787 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272648 4787 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272653 4787 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272657 4787 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272662 4787 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272668 4787 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272672 4787 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272677 4787 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272681 4787 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272686 4787 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272690 4787 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272695 4787 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272699 4787 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272704 4787 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272708 4787 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272713 4787 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272717 4787 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272721 4787 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272726 4787 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272731 4787 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272735 4787 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272740 4787 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272746 4787 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272776 4787 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272782 4787 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272788 4787 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272795 4787 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272801 4787 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272806 4787 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272811 4787 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272816 4787 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272821 4787 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272827 4787 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272832 4787 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272837 4787 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272842 4787 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272848 4787 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272852 4787 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272856 4787 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272861 4787 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272865 4787 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272871 4787 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272877 4787 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272881 4787 feature_gate.go:330] unrecognized feature gate: Example Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272886 4787 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272891 4787 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272895 4787 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272902 4787 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272908 4787 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272913 4787 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272918 4787 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272923 4787 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272929 4787 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272934 4787 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272938 4787 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272943 4787 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272947 4787 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272952 4787 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272957 4787 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272963 4787 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272968 4787 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272972 4787 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272977 4787 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.272981 4787 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273099 4787 flags.go:64] FLAG: --address="0.0.0.0" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273113 4787 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273123 4787 flags.go:64] FLAG: --anonymous-auth="true" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273131 4787 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273139 4787 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273145 4787 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273153 4787 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273161 4787 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273167 4787 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273173 4787 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273179 4787 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273185 4787 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273190 4787 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273196 4787 flags.go:64] FLAG: --cgroup-root="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273201 4787 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273206 4787 flags.go:64] FLAG: --client-ca-file="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273212 4787 flags.go:64] FLAG: --cloud-config="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273217 4787 flags.go:64] FLAG: --cloud-provider="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273223 4787 flags.go:64] FLAG: --cluster-dns="[]" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273233 4787 flags.go:64] FLAG: --cluster-domain="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273238 4787 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273245 4787 flags.go:64] FLAG: --config-dir="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273251 4787 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273257 4787 flags.go:64] FLAG: --container-log-max-files="5" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273265 4787 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273271 4787 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273276 4787 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273283 4787 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273288 4787 flags.go:64] FLAG: --contention-profiling="false" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273293 4787 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273299 4787 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273304 4787 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273310 4787 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273317 4787 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273322 4787 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273328 4787 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273333 4787 flags.go:64] FLAG: --enable-load-reader="false" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273338 4787 flags.go:64] FLAG: --enable-server="true" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273344 4787 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273373 4787 flags.go:64] FLAG: --event-burst="100" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273379 4787 flags.go:64] FLAG: --event-qps="50" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273385 4787 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273390 4787 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273395 4787 flags.go:64] FLAG: --eviction-hard="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273410 4787 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273415 4787 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273420 4787 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273426 4787 flags.go:64] FLAG: --eviction-soft="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273432 4787 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273437 4787 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273442 4787 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273450 4787 flags.go:64] FLAG: --experimental-mounter-path="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273455 4787 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273461 4787 flags.go:64] FLAG: --fail-swap-on="true" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273467 4787 flags.go:64] FLAG: --feature-gates="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273474 4787 flags.go:64] FLAG: --file-check-frequency="20s" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273480 4787 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273486 4787 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273491 4787 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273497 4787 flags.go:64] FLAG: --healthz-port="10248" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273503 4787 flags.go:64] FLAG: --help="false" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273508 4787 flags.go:64] FLAG: --hostname-override="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273513 4787 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273519 4787 flags.go:64] FLAG: --http-check-frequency="20s" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273525 4787 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273530 4787 flags.go:64] FLAG: --image-credential-provider-config="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273535 4787 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273541 4787 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273546 4787 flags.go:64] FLAG: --image-service-endpoint="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273551 4787 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273556 4787 flags.go:64] FLAG: --kube-api-burst="100" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273561 4787 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273567 4787 flags.go:64] FLAG: --kube-api-qps="50" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273571 4787 flags.go:64] FLAG: --kube-reserved="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273576 4787 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273581 4787 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273586 4787 flags.go:64] FLAG: --kubelet-cgroups="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273591 4787 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273597 4787 flags.go:64] FLAG: --lock-file="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273601 4787 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273608 4787 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273614 4787 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273622 4787 flags.go:64] FLAG: --log-json-split-stream="false" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273628 4787 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273633 4787 flags.go:64] FLAG: --log-text-split-stream="false" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273639 4787 flags.go:64] FLAG: --logging-format="text" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273644 4787 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273650 4787 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273655 4787 flags.go:64] FLAG: --manifest-url="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273660 4787 flags.go:64] FLAG: --manifest-url-header="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273668 4787 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273673 4787 flags.go:64] FLAG: --max-open-files="1000000" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273680 4787 flags.go:64] FLAG: --max-pods="110" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273685 4787 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273691 4787 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273696 4787 flags.go:64] FLAG: --memory-manager-policy="None" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273702 4787 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273708 4787 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273714 4787 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273719 4787 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273733 4787 flags.go:64] FLAG: --node-status-max-images="50" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273739 4787 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273744 4787 flags.go:64] FLAG: --oom-score-adj="-999" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273750 4787 flags.go:64] FLAG: --pod-cidr="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273755 4787 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273763 4787 flags.go:64] FLAG: --pod-manifest-path="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273769 4787 flags.go:64] FLAG: --pod-max-pids="-1" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273774 4787 flags.go:64] FLAG: --pods-per-core="0" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273779 4787 flags.go:64] FLAG: --port="10250" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273784 4787 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273790 4787 flags.go:64] FLAG: --provider-id="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273795 4787 flags.go:64] FLAG: --qos-reserved="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273801 4787 flags.go:64] FLAG: --read-only-port="10255" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273806 4787 flags.go:64] FLAG: --register-node="true" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273812 4787 flags.go:64] FLAG: --register-schedulable="true" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273816 4787 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273827 4787 flags.go:64] FLAG: --registry-burst="10" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273833 4787 flags.go:64] FLAG: --registry-qps="5" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273838 4787 flags.go:64] FLAG: --reserved-cpus="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273843 4787 flags.go:64] FLAG: --reserved-memory="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273850 4787 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273856 4787 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273861 4787 flags.go:64] FLAG: --rotate-certificates="false" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273866 4787 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273871 4787 flags.go:64] FLAG: --runonce="false" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273876 4787 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273881 4787 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273886 4787 flags.go:64] FLAG: --seccomp-default="false" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273892 4787 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273897 4787 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273903 4787 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273908 4787 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273913 4787 flags.go:64] FLAG: --storage-driver-password="root" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273918 4787 flags.go:64] FLAG: --storage-driver-secure="false" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273924 4787 flags.go:64] FLAG: --storage-driver-table="stats" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273929 4787 flags.go:64] FLAG: --storage-driver-user="root" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273935 4787 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273941 4787 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273947 4787 flags.go:64] FLAG: --system-cgroups="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273953 4787 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273962 4787 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273967 4787 flags.go:64] FLAG: --tls-cert-file="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273973 4787 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273980 4787 flags.go:64] FLAG: --tls-min-version="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273987 4787 flags.go:64] FLAG: --tls-private-key-file="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273992 4787 flags.go:64] FLAG: --topology-manager-policy="none" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.273997 4787 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.274003 4787 flags.go:64] FLAG: --topology-manager-scope="container" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.274009 4787 flags.go:64] FLAG: --v="2" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.274016 4787 flags.go:64] FLAG: --version="false" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.274023 4787 flags.go:64] FLAG: --vmodule="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.274030 4787 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.274037 4787 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274180 4787 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274188 4787 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274194 4787 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274200 4787 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274206 4787 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274210 4787 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274215 4787 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274219 4787 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274224 4787 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274229 4787 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274233 4787 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274237 4787 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274242 4787 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274246 4787 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274250 4787 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274256 4787 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274262 4787 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274267 4787 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274272 4787 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274276 4787 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274281 4787 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274285 4787 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274290 4787 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274295 4787 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274300 4787 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274305 4787 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274310 4787 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274314 4787 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274319 4787 feature_gate.go:330] unrecognized feature gate: Example Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274324 4787 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274328 4787 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274333 4787 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274338 4787 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274342 4787 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274346 4787 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274352 4787 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274357 4787 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274362 4787 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274366 4787 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274372 4787 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274376 4787 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274381 4787 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274385 4787 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274390 4787 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274394 4787 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274399 4787 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274403 4787 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274408 4787 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274412 4787 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274417 4787 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274421 4787 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274425 4787 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274430 4787 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274435 4787 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274440 4787 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274444 4787 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274450 4787 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274455 4787 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274460 4787 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274464 4787 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274469 4787 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274474 4787 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274479 4787 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274483 4787 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274490 4787 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274496 4787 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274502 4787 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274506 4787 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274511 4787 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274516 4787 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.274521 4787 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.274532 4787 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.287930 4787 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.287993 4787 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288178 4787 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288209 4787 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288224 4787 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288236 4787 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288250 4787 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288260 4787 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288269 4787 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288277 4787 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288285 4787 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288293 4787 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288302 4787 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288313 4787 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288325 4787 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288337 4787 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288347 4787 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288357 4787 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288365 4787 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288375 4787 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288385 4787 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288395 4787 feature_gate.go:330] unrecognized feature gate: Example Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288405 4787 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288416 4787 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288426 4787 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288434 4787 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288444 4787 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288454 4787 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288463 4787 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288471 4787 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288480 4787 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288487 4787 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288495 4787 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288503 4787 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288511 4787 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288518 4787 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288529 4787 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288537 4787 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288545 4787 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288553 4787 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288561 4787 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288569 4787 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288577 4787 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288585 4787 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288593 4787 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288602 4787 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288611 4787 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288622 4787 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288632 4787 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288642 4787 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288652 4787 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288661 4787 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288671 4787 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288682 4787 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288695 4787 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288705 4787 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288714 4787 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288723 4787 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288731 4787 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288739 4787 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288747 4787 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288755 4787 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288762 4787 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288770 4787 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288778 4787 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288786 4787 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288795 4787 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288802 4787 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288810 4787 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288818 4787 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288826 4787 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288833 4787 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.288842 4787 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.288855 4787 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289121 4787 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289137 4787 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289148 4787 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289157 4787 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289165 4787 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289174 4787 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289182 4787 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289189 4787 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289197 4787 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289206 4787 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289217 4787 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289229 4787 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289238 4787 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289246 4787 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289254 4787 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289262 4787 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289270 4787 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289278 4787 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289286 4787 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289294 4787 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289303 4787 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289313 4787 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289323 4787 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289332 4787 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289342 4787 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289350 4787 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289358 4787 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289367 4787 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289377 4787 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289387 4787 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289395 4787 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289404 4787 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289413 4787 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289423 4787 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289435 4787 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289445 4787 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289455 4787 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289484 4787 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289493 4787 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289503 4787 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289513 4787 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289523 4787 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289533 4787 feature_gate.go:330] unrecognized feature gate: Example Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289543 4787 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289553 4787 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289565 4787 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289574 4787 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289583 4787 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289593 4787 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289603 4787 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289613 4787 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289622 4787 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289633 4787 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289643 4787 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289651 4787 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289661 4787 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289672 4787 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289680 4787 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289688 4787 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289696 4787 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289704 4787 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289712 4787 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289720 4787 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289727 4787 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289735 4787 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289743 4787 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289750 4787 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289758 4787 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289766 4787 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289774 4787 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.289785 4787 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.289800 4787 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.290841 4787 server.go:940] "Client rotation is on, will bootstrap in background" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.295843 4787 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.295940 4787 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.297125 4787 server.go:997] "Starting client certificate rotation" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.297147 4787 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.298146 4787 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-29 01:06:37.47370457 +0000 UTC Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.298266 4787 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 2127h30m29.175443495s for next certificate rotation Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.324350 4787 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.326168 4787 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.350945 4787 log.go:25] "Validated CRI v1 runtime API" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.389184 4787 log.go:25] "Validated CRI v1 image API" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.391270 4787 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.396423 4787 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-01-09-00-28-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.396484 4787 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.429397 4787 manager.go:217] Machine: {Timestamp:2025-10-01 09:36:08.425851981 +0000 UTC m=+0.540996228 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:869e21e0-7910-4427-aeba-5df675f5c14f BootID:7f63d222-e40c-4123-900b-74cfec920eac Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:e0:41:9d Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:e0:41:9d Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:96:51:73 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:5c:66:fd Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:df:3b:b9 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:1d:74:df Speed:-1 Mtu:1496} {Name:eth10 MacAddress:ae:96:de:4c:ec:52 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:fe:c0:63:f2:ba:14 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.429772 4787 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.429987 4787 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.430530 4787 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.430883 4787 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.430974 4787 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.431419 4787 topology_manager.go:138] "Creating topology manager with none policy" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.431439 4787 container_manager_linux.go:303] "Creating device plugin manager" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.431985 4787 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.432033 4787 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.432342 4787 state_mem.go:36] "Initialized new in-memory state store" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.432496 4787 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.436407 4787 kubelet.go:418] "Attempting to sync node with API server" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.436440 4787 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.436517 4787 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.436549 4787 kubelet.go:324] "Adding apiserver pod source" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.436572 4787 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.441347 4787 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.442469 4787 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.443914 4787 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.39:6443: connect: connection refused Oct 01 09:36:08 crc kubenswrapper[4787]: E1001 09:36:08.444072 4787 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.39:6443: connect: connection refused" logger="UnhandledError" Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.444469 4787 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.39:6443: connect: connection refused Oct 01 09:36:08 crc kubenswrapper[4787]: E1001 09:36:08.444555 4787 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.39:6443: connect: connection refused" logger="UnhandledError" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.445984 4787 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.447636 4787 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.447682 4787 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.447703 4787 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.447721 4787 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.447750 4787 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.447768 4787 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.447786 4787 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.447813 4787 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.447829 4787 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.447845 4787 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.447862 4787 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.447876 4787 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.449896 4787 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.450553 4787 server.go:1280] "Started kubelet" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.454117 4787 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.454036 4787 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.451289 4787 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.39:6443: connect: connection refused Oct 01 09:36:08 crc systemd[1]: Started Kubernetes Kubelet. Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.456041 4787 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.458977 4787 server.go:460] "Adding debug handlers to kubelet server" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.459135 4787 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.459327 4787 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 09:53:47.931376873 +0000 UTC Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.459369 4787 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1968h17m39.472011201s for next certificate rotation Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.459598 4787 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.459775 4787 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.459801 4787 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 01 09:36:08 crc kubenswrapper[4787]: E1001 09:36:08.468223 4787 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.469228 4787 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 01 09:36:08 crc kubenswrapper[4787]: E1001 09:36:08.468559 4787 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.129.56.39:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186a545e4dcc8bb2 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-01 09:36:08.450509746 +0000 UTC m=+0.565653943,LastTimestamp:2025-10-01 09:36:08.450509746 +0000 UTC m=+0.565653943,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.469662 4787 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.39:6443: connect: connection refused Oct 01 09:36:08 crc kubenswrapper[4787]: E1001 09:36:08.469764 4787 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.39:6443: connect: connection refused" logger="UnhandledError" Oct 01 09:36:08 crc kubenswrapper[4787]: E1001 09:36:08.469398 4787 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.39:6443: connect: connection refused" interval="200ms" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.471155 4787 factory.go:153] Registering CRI-O factory Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.471182 4787 factory.go:221] Registration of the crio container factory successfully Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.471248 4787 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.471259 4787 factory.go:55] Registering systemd factory Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.471267 4787 factory.go:221] Registration of the systemd container factory successfully Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.471299 4787 factory.go:103] Registering Raw factory Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.471318 4787 manager.go:1196] Started watching for new ooms in manager Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.471896 4787 manager.go:319] Starting recovery of all containers Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.472652 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.472694 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.472708 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.472720 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.472732 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.472743 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.472755 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.472766 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.472780 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.472793 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.472810 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.472822 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.472834 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.472847 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.472890 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.472903 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.472916 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.472927 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.472940 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.472952 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.472989 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473000 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473013 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473024 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473035 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473046 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473062 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473120 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473134 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473147 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473159 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473172 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473183 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473194 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473206 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473218 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473229 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473240 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473251 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473263 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473274 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473286 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473307 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473322 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473334 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473346 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473358 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473396 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473411 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473423 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473434 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473470 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473488 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473502 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473516 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473528 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473540 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473553 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473564 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473576 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473589 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473600 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473611 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473622 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473634 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473644 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473656 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473667 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473680 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473691 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473702 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473713 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473725 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473736 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473747 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473757 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473767 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473779 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473790 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473801 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473809 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473817 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473825 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473833 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473841 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473850 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473859 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473867 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473875 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473883 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473891 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473900 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473908 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473916 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473925 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.473933 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.475754 4787 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.475787 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.475804 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.475818 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.475831 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.475848 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.475861 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.475875 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.475888 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.475907 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.475928 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.475943 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.475957 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.475969 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.475984 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.475996 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.476009 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.476021 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478304 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478327 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478336 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478345 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478353 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478361 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478370 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478379 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478388 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478429 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478438 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478446 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478455 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478463 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478472 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478481 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478490 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478498 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478507 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478515 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478523 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478533 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478543 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478552 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478560 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478569 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478577 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478586 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478594 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478603 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478615 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478627 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478639 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478652 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478661 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478669 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478676 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478684 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478692 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478701 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478710 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478720 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478728 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478737 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478746 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478755 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478764 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478773 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478782 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478792 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478800 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478810 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478818 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478828 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478837 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478845 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478854 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478863 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478872 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478881 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478890 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478934 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478944 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478953 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478961 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478973 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478983 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.478992 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.479001 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.479011 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.479021 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.479030 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.479040 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.479050 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.479058 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.479068 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.479093 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.479104 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.479113 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.479122 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.479131 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.479141 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.479151 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.479161 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.479170 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.479180 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.479189 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.479198 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.479207 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.479217 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.479226 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.479234 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.479243 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.479256 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.479264 4787 reconstruct.go:97] "Volume reconstruction finished" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.479271 4787 reconciler.go:26] "Reconciler: start to sync state" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.492232 4787 manager.go:324] Recovery completed Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.502960 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.505331 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.505367 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.505376 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.506101 4787 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.506112 4787 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.506150 4787 state_mem.go:36] "Initialized new in-memory state store" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.518939 4787 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.521749 4787 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.522032 4787 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.522355 4787 kubelet.go:2335] "Starting kubelet main sync loop" Oct 01 09:36:08 crc kubenswrapper[4787]: E1001 09:36:08.522822 4787 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 01 09:36:08 crc kubenswrapper[4787]: W1001 09:36:08.523172 4787 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.39:6443: connect: connection refused Oct 01 09:36:08 crc kubenswrapper[4787]: E1001 09:36:08.523366 4787 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.39:6443: connect: connection refused" logger="UnhandledError" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.528042 4787 policy_none.go:49] "None policy: Start" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.528771 4787 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.528793 4787 state_mem.go:35] "Initializing new in-memory state store" Oct 01 09:36:08 crc kubenswrapper[4787]: E1001 09:36:08.568921 4787 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.601381 4787 manager.go:334] "Starting Device Plugin manager" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.601442 4787 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.601458 4787 server.go:79] "Starting device plugin registration server" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.601984 4787 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.602001 4787 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.602288 4787 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.602371 4787 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.602380 4787 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 01 09:36:08 crc kubenswrapper[4787]: E1001 09:36:08.610885 4787 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.623705 4787 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc"] Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.623809 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.627315 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.627345 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.627354 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.627455 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.627867 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.627944 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.628169 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.628191 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.628199 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.628262 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.628376 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.628403 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.629013 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.629033 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.629041 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.629158 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.629416 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.629463 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.629487 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.629514 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.629514 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.629761 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.629778 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.629786 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.629859 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.630040 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.630135 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.631477 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.631588 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.631605 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.631878 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.631928 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.632489 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.632509 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.632516 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.632580 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.632605 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.632640 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.632666 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.632682 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.632827 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.632842 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.632849 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.632648 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:08 crc kubenswrapper[4787]: E1001 09:36:08.670618 4787 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.39:6443: connect: connection refused" interval="400ms" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.682012 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.682200 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.682233 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.682257 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.682278 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.682295 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.682317 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.682430 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.682460 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.682482 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.682503 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.682524 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.682546 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.682573 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.682596 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.703047 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.703929 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.703956 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.703964 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.703980 4787 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 01 09:36:08 crc kubenswrapper[4787]: E1001 09:36:08.704230 4787 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.39:6443: connect: connection refused" node="crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.783026 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.783111 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.783134 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.783156 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.783180 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.783182 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.783240 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.783266 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.783274 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.783284 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.783202 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.783326 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.783328 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.783345 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.783365 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.783413 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.783450 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.783477 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.783499 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.783521 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.783540 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.783621 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.783641 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.783562 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.783593 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.783576 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.783594 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.783743 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.783759 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.783579 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.904533 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.905593 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.905617 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.905626 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.905646 4787 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 01 09:36:08 crc kubenswrapper[4787]: E1001 09:36:08.905867 4787 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.39:6443: connect: connection refused" node="crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.966760 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 09:36:08 crc kubenswrapper[4787]: I1001 09:36:08.983476 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 09:36:09 crc kubenswrapper[4787]: I1001 09:36:09.000766 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 01 09:36:09 crc kubenswrapper[4787]: I1001 09:36:09.010951 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 09:36:09 crc kubenswrapper[4787]: I1001 09:36:09.016502 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 01 09:36:09 crc kubenswrapper[4787]: W1001 09:36:09.025938 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-5cfd18f90fdbeea502804d2d42ed5a7ed73d892c025ac4f9febc68945d39017c WatchSource:0}: Error finding container 5cfd18f90fdbeea502804d2d42ed5a7ed73d892c025ac4f9febc68945d39017c: Status 404 returned error can't find the container with id 5cfd18f90fdbeea502804d2d42ed5a7ed73d892c025ac4f9febc68945d39017c Oct 01 09:36:09 crc kubenswrapper[4787]: W1001 09:36:09.032207 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-d4d8dedd4993430393c33db9d40acd4f481caa83100318b9557bc1c0d51d3386 WatchSource:0}: Error finding container d4d8dedd4993430393c33db9d40acd4f481caa83100318b9557bc1c0d51d3386: Status 404 returned error can't find the container with id d4d8dedd4993430393c33db9d40acd4f481caa83100318b9557bc1c0d51d3386 Oct 01 09:36:09 crc kubenswrapper[4787]: W1001 09:36:09.045615 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-d188c5a8e8a8131a0d956a56bb430839b8525276be384fc9937c53199f9ba840 WatchSource:0}: Error finding container d188c5a8e8a8131a0d956a56bb430839b8525276be384fc9937c53199f9ba840: Status 404 returned error can't find the container with id d188c5a8e8a8131a0d956a56bb430839b8525276be384fc9937c53199f9ba840 Oct 01 09:36:09 crc kubenswrapper[4787]: W1001 09:36:09.047663 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-4f9b5d409d620e640b526450e6a61bda8c5fec555c25f3a508650d488181e561 WatchSource:0}: Error finding container 4f9b5d409d620e640b526450e6a61bda8c5fec555c25f3a508650d488181e561: Status 404 returned error can't find the container with id 4f9b5d409d620e640b526450e6a61bda8c5fec555c25f3a508650d488181e561 Oct 01 09:36:09 crc kubenswrapper[4787]: W1001 09:36:09.049598 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-c6be33b011a77b09daabc0380c1e14469d443241c1ed3bad1e82a69efa787eee WatchSource:0}: Error finding container c6be33b011a77b09daabc0380c1e14469d443241c1ed3bad1e82a69efa787eee: Status 404 returned error can't find the container with id c6be33b011a77b09daabc0380c1e14469d443241c1ed3bad1e82a69efa787eee Oct 01 09:36:09 crc kubenswrapper[4787]: E1001 09:36:09.071496 4787 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.39:6443: connect: connection refused" interval="800ms" Oct 01 09:36:09 crc kubenswrapper[4787]: I1001 09:36:09.307006 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:09 crc kubenswrapper[4787]: I1001 09:36:09.308794 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:09 crc kubenswrapper[4787]: I1001 09:36:09.308855 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:09 crc kubenswrapper[4787]: I1001 09:36:09.308880 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:09 crc kubenswrapper[4787]: I1001 09:36:09.308929 4787 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 01 09:36:09 crc kubenswrapper[4787]: E1001 09:36:09.309643 4787 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.39:6443: connect: connection refused" node="crc" Oct 01 09:36:09 crc kubenswrapper[4787]: I1001 09:36:09.456017 4787 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.39:6443: connect: connection refused Oct 01 09:36:09 crc kubenswrapper[4787]: W1001 09:36:09.504296 4787 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.39:6443: connect: connection refused Oct 01 09:36:09 crc kubenswrapper[4787]: E1001 09:36:09.504378 4787 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.39:6443: connect: connection refused" logger="UnhandledError" Oct 01 09:36:09 crc kubenswrapper[4787]: I1001 09:36:09.526997 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"4f9b5d409d620e640b526450e6a61bda8c5fec555c25f3a508650d488181e561"} Oct 01 09:36:09 crc kubenswrapper[4787]: I1001 09:36:09.528034 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c6be33b011a77b09daabc0380c1e14469d443241c1ed3bad1e82a69efa787eee"} Oct 01 09:36:09 crc kubenswrapper[4787]: I1001 09:36:09.529476 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d188c5a8e8a8131a0d956a56bb430839b8525276be384fc9937c53199f9ba840"} Oct 01 09:36:09 crc kubenswrapper[4787]: I1001 09:36:09.530694 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5cfd18f90fdbeea502804d2d42ed5a7ed73d892c025ac4f9febc68945d39017c"} Oct 01 09:36:09 crc kubenswrapper[4787]: I1001 09:36:09.532776 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"d4d8dedd4993430393c33db9d40acd4f481caa83100318b9557bc1c0d51d3386"} Oct 01 09:36:09 crc kubenswrapper[4787]: W1001 09:36:09.609470 4787 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.39:6443: connect: connection refused Oct 01 09:36:09 crc kubenswrapper[4787]: E1001 09:36:09.609882 4787 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.39:6443: connect: connection refused" logger="UnhandledError" Oct 01 09:36:09 crc kubenswrapper[4787]: W1001 09:36:09.704593 4787 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.39:6443: connect: connection refused Oct 01 09:36:09 crc kubenswrapper[4787]: E1001 09:36:09.704665 4787 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.39:6443: connect: connection refused" logger="UnhandledError" Oct 01 09:36:09 crc kubenswrapper[4787]: E1001 09:36:09.872922 4787 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.39:6443: connect: connection refused" interval="1.6s" Oct 01 09:36:10 crc kubenswrapper[4787]: W1001 09:36:10.037743 4787 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.39:6443: connect: connection refused Oct 01 09:36:10 crc kubenswrapper[4787]: E1001 09:36:10.037851 4787 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.39:6443: connect: connection refused" logger="UnhandledError" Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.110190 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.111487 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.111535 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.111549 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.111574 4787 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 01 09:36:10 crc kubenswrapper[4787]: E1001 09:36:10.112019 4787 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.39:6443: connect: connection refused" node="crc" Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.455270 4787 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.39:6443: connect: connection refused Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.537669 4787 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="9b08e51525e5ac6c978ce0e0634e37e466712e2674a54654851617cab894d2e0" exitCode=0 Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.537741 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"9b08e51525e5ac6c978ce0e0634e37e466712e2674a54654851617cab894d2e0"} Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.537862 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.538643 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.538664 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.538673 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.547011 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"144123d273f3f1af4fb857d731a16d9891b097c98229a035eb132f360fa8b2f4"} Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.547095 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.547115 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c8d4c78dd607b81a28bcfe0ffa69c02628cf06aff5fd6a574069639c5bac044a"} Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.547132 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a1bb16edb25b63e957cc0bceb04a9fa9244e68881cb45be7c7d627c5faa74232"} Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.547144 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"995030421e0cbfeb7baa8ccb574d269dd178f8185bcaa9000fe5e7fc221b1188"} Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.548998 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.549068 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.549118 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.552350 4787 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="7c3cd7c1e19e00dbc9cb8cb8f3f857ae636ff3818c589769f15aa87cb94327ce" exitCode=0 Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.552424 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.552418 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"7c3cd7c1e19e00dbc9cb8cb8f3f857ae636ff3818c589769f15aa87cb94327ce"} Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.553231 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.553268 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.553282 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.554646 4787 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="5d5cf849bb15a6e2c7abb11fc455dd38be5e20a59158454f27ef69541ff50765" exitCode=0 Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.554701 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.554698 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"5d5cf849bb15a6e2c7abb11fc455dd38be5e20a59158454f27ef69541ff50765"} Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.556174 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.556228 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.556246 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.556893 4787 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e" exitCode=0 Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.556936 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e"} Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.557040 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.561511 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.561553 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.561571 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.566466 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.570216 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.570254 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:10 crc kubenswrapper[4787]: I1001 09:36:10.570269 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:11 crc kubenswrapper[4787]: W1001 09:36:11.391796 4787 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.39:6443: connect: connection refused Oct 01 09:36:11 crc kubenswrapper[4787]: E1001 09:36:11.391918 4787 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.39:6443: connect: connection refused" logger="UnhandledError" Oct 01 09:36:11 crc kubenswrapper[4787]: I1001 09:36:11.455731 4787 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.39:6443: connect: connection refused Oct 01 09:36:11 crc kubenswrapper[4787]: E1001 09:36:11.474461 4787 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.39:6443: connect: connection refused" interval="3.2s" Oct 01 09:36:11 crc kubenswrapper[4787]: W1001 09:36:11.500046 4787 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.39:6443: connect: connection refused Oct 01 09:36:11 crc kubenswrapper[4787]: E1001 09:36:11.500148 4787 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.39:6443: connect: connection refused" logger="UnhandledError" Oct 01 09:36:11 crc kubenswrapper[4787]: I1001 09:36:11.562527 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e1d95bbc5f966defa47192b7cc53c32f75894f4b51ce8411d7bf828e04c282b9"} Oct 01 09:36:11 crc kubenswrapper[4787]: I1001 09:36:11.562570 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"4f63e01c8f45be048cd6ade60cb51f054fbea42bb51d7b5cd32f71cf1f15ce21"} Oct 01 09:36:11 crc kubenswrapper[4787]: I1001 09:36:11.562580 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"f445035e4b6faf9fca8e0dbe8c69668de8825ccebd6ace48f80685c88312b707"} Oct 01 09:36:11 crc kubenswrapper[4787]: I1001 09:36:11.562590 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:11 crc kubenswrapper[4787]: I1001 09:36:11.563451 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:11 crc kubenswrapper[4787]: I1001 09:36:11.563494 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:11 crc kubenswrapper[4787]: I1001 09:36:11.563508 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:11 crc kubenswrapper[4787]: I1001 09:36:11.564311 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:11 crc kubenswrapper[4787]: I1001 09:36:11.564306 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"93d6842d7f8928e5a2b557a70568f5b895e17f9be42d3fadae0b91ba692f5fd9"} Oct 01 09:36:11 crc kubenswrapper[4787]: I1001 09:36:11.564977 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:11 crc kubenswrapper[4787]: I1001 09:36:11.565011 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:11 crc kubenswrapper[4787]: I1001 09:36:11.565021 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:11 crc kubenswrapper[4787]: I1001 09:36:11.567702 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c"} Oct 01 09:36:11 crc kubenswrapper[4787]: I1001 09:36:11.567733 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:11 crc kubenswrapper[4787]: I1001 09:36:11.567733 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730"} Oct 01 09:36:11 crc kubenswrapper[4787]: I1001 09:36:11.567861 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644"} Oct 01 09:36:11 crc kubenswrapper[4787]: I1001 09:36:11.567881 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1"} Oct 01 09:36:11 crc kubenswrapper[4787]: I1001 09:36:11.567894 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283"} Oct 01 09:36:11 crc kubenswrapper[4787]: I1001 09:36:11.568450 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:11 crc kubenswrapper[4787]: I1001 09:36:11.568477 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:11 crc kubenswrapper[4787]: I1001 09:36:11.568492 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:11 crc kubenswrapper[4787]: I1001 09:36:11.573134 4787 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="e8624a723122aebef9d2845bfabbd2cdecd98c8e970e511d1a37ce1eba8c19a5" exitCode=0 Oct 01 09:36:11 crc kubenswrapper[4787]: I1001 09:36:11.573302 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:11 crc kubenswrapper[4787]: I1001 09:36:11.573501 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"e8624a723122aebef9d2845bfabbd2cdecd98c8e970e511d1a37ce1eba8c19a5"} Oct 01 09:36:11 crc kubenswrapper[4787]: I1001 09:36:11.573525 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:11 crc kubenswrapper[4787]: I1001 09:36:11.574204 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:11 crc kubenswrapper[4787]: I1001 09:36:11.574239 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:11 crc kubenswrapper[4787]: I1001 09:36:11.574251 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:11 crc kubenswrapper[4787]: I1001 09:36:11.574979 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:11 crc kubenswrapper[4787]: I1001 09:36:11.575002 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:11 crc kubenswrapper[4787]: I1001 09:36:11.575012 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:11 crc kubenswrapper[4787]: I1001 09:36:11.712640 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:11 crc kubenswrapper[4787]: I1001 09:36:11.713842 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:11 crc kubenswrapper[4787]: I1001 09:36:11.713869 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:11 crc kubenswrapper[4787]: I1001 09:36:11.713877 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:11 crc kubenswrapper[4787]: I1001 09:36:11.713916 4787 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 01 09:36:11 crc kubenswrapper[4787]: E1001 09:36:11.714460 4787 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.39:6443: connect: connection refused" node="crc" Oct 01 09:36:11 crc kubenswrapper[4787]: W1001 09:36:11.822925 4787 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.39:6443: connect: connection refused Oct 01 09:36:11 crc kubenswrapper[4787]: E1001 09:36:11.823120 4787 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.39:6443: connect: connection refused" logger="UnhandledError" Oct 01 09:36:12 crc kubenswrapper[4787]: I1001 09:36:12.577619 4787 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="c847455912fab5587f8318390eca0cd83c4c779228795ac9804569b78cdaa2e2" exitCode=0 Oct 01 09:36:12 crc kubenswrapper[4787]: I1001 09:36:12.577688 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"c847455912fab5587f8318390eca0cd83c4c779228795ac9804569b78cdaa2e2"} Oct 01 09:36:12 crc kubenswrapper[4787]: I1001 09:36:12.577719 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:12 crc kubenswrapper[4787]: I1001 09:36:12.577732 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:12 crc kubenswrapper[4787]: I1001 09:36:12.577822 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:12 crc kubenswrapper[4787]: I1001 09:36:12.577923 4787 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 01 09:36:12 crc kubenswrapper[4787]: I1001 09:36:12.578025 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:12 crc kubenswrapper[4787]: I1001 09:36:12.578531 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 09:36:12 crc kubenswrapper[4787]: I1001 09:36:12.578862 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:12 crc kubenswrapper[4787]: I1001 09:36:12.578880 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:12 crc kubenswrapper[4787]: I1001 09:36:12.578889 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:12 crc kubenswrapper[4787]: I1001 09:36:12.578896 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:12 crc kubenswrapper[4787]: I1001 09:36:12.578909 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:12 crc kubenswrapper[4787]: I1001 09:36:12.578921 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:12 crc kubenswrapper[4787]: I1001 09:36:12.578899 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:12 crc kubenswrapper[4787]: I1001 09:36:12.579019 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:12 crc kubenswrapper[4787]: I1001 09:36:12.578910 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:12 crc kubenswrapper[4787]: I1001 09:36:12.579548 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:12 crc kubenswrapper[4787]: I1001 09:36:12.579600 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:12 crc kubenswrapper[4787]: I1001 09:36:12.579613 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:13 crc kubenswrapper[4787]: I1001 09:36:13.206767 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 09:36:13 crc kubenswrapper[4787]: I1001 09:36:13.207023 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:13 crc kubenswrapper[4787]: I1001 09:36:13.208634 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:13 crc kubenswrapper[4787]: I1001 09:36:13.208699 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:13 crc kubenswrapper[4787]: I1001 09:36:13.208719 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:13 crc kubenswrapper[4787]: I1001 09:36:13.585748 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ab0ed2f4328676868b7e254ce30122ea0b9188c1995ed2d3d9a3366a275bd0a0"} Oct 01 09:36:13 crc kubenswrapper[4787]: I1001 09:36:13.585790 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"09e24046b37ecd75dc626bb5615f63b55ae7cfff90b989d2cbd6d7c4c0c13c02"} Oct 01 09:36:13 crc kubenswrapper[4787]: I1001 09:36:13.585804 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"0dda23a8b4d17dbb28895da4f9529035351a75c71f40ed4d6e63203ab7f40d8f"} Oct 01 09:36:13 crc kubenswrapper[4787]: I1001 09:36:13.585814 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:13 crc kubenswrapper[4787]: I1001 09:36:13.585815 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"04a96ba027ceb6bf419cb549f3d63d95a871dd08b3ce4165613a721903127d8a"} Oct 01 09:36:13 crc kubenswrapper[4787]: I1001 09:36:13.587300 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:13 crc kubenswrapper[4787]: I1001 09:36:13.587421 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:13 crc kubenswrapper[4787]: I1001 09:36:13.587454 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:13 crc kubenswrapper[4787]: I1001 09:36:13.948143 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 09:36:13 crc kubenswrapper[4787]: I1001 09:36:13.948360 4787 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 01 09:36:13 crc kubenswrapper[4787]: I1001 09:36:13.948426 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:13 crc kubenswrapper[4787]: I1001 09:36:13.953072 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:13 crc kubenswrapper[4787]: I1001 09:36:13.953155 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:13 crc kubenswrapper[4787]: I1001 09:36:13.953171 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:14 crc kubenswrapper[4787]: I1001 09:36:14.591853 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5b812bf26eb34ca5983637e0ce89cd9fd38fb8c5d7c1eac26ab7c76a5dafa876"} Oct 01 09:36:14 crc kubenswrapper[4787]: I1001 09:36:14.591999 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:14 crc kubenswrapper[4787]: I1001 09:36:14.593066 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:14 crc kubenswrapper[4787]: I1001 09:36:14.593118 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:14 crc kubenswrapper[4787]: I1001 09:36:14.593131 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:14 crc kubenswrapper[4787]: I1001 09:36:14.914919 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:14 crc kubenswrapper[4787]: I1001 09:36:14.916859 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:14 crc kubenswrapper[4787]: I1001 09:36:14.916899 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:14 crc kubenswrapper[4787]: I1001 09:36:14.916908 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:14 crc kubenswrapper[4787]: I1001 09:36:14.916932 4787 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 01 09:36:15 crc kubenswrapper[4787]: I1001 09:36:15.315332 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 09:36:15 crc kubenswrapper[4787]: I1001 09:36:15.315657 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:15 crc kubenswrapper[4787]: I1001 09:36:15.317042 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:15 crc kubenswrapper[4787]: I1001 09:36:15.317089 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:15 crc kubenswrapper[4787]: I1001 09:36:15.317099 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:15 crc kubenswrapper[4787]: I1001 09:36:15.409443 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 01 09:36:15 crc kubenswrapper[4787]: I1001 09:36:15.588911 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 01 09:36:15 crc kubenswrapper[4787]: I1001 09:36:15.593860 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:15 crc kubenswrapper[4787]: I1001 09:36:15.594700 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:15 crc kubenswrapper[4787]: I1001 09:36:15.594758 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:15 crc kubenswrapper[4787]: I1001 09:36:15.594775 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:16 crc kubenswrapper[4787]: I1001 09:36:16.043465 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 09:36:16 crc kubenswrapper[4787]: I1001 09:36:16.043697 4787 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 01 09:36:16 crc kubenswrapper[4787]: I1001 09:36:16.043756 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:16 crc kubenswrapper[4787]: I1001 09:36:16.045509 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:16 crc kubenswrapper[4787]: I1001 09:36:16.045608 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:16 crc kubenswrapper[4787]: I1001 09:36:16.045636 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:16 crc kubenswrapper[4787]: I1001 09:36:16.207579 4787 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 01 09:36:16 crc kubenswrapper[4787]: I1001 09:36:16.207690 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 01 09:36:16 crc kubenswrapper[4787]: I1001 09:36:16.596227 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:16 crc kubenswrapper[4787]: I1001 09:36:16.597969 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:16 crc kubenswrapper[4787]: I1001 09:36:16.598029 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:16 crc kubenswrapper[4787]: I1001 09:36:16.598049 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:17 crc kubenswrapper[4787]: I1001 09:36:17.091755 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 09:36:17 crc kubenswrapper[4787]: I1001 09:36:17.091925 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:17 crc kubenswrapper[4787]: I1001 09:36:17.093039 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:17 crc kubenswrapper[4787]: I1001 09:36:17.093124 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:17 crc kubenswrapper[4787]: I1001 09:36:17.093141 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:17 crc kubenswrapper[4787]: I1001 09:36:17.097723 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 09:36:17 crc kubenswrapper[4787]: I1001 09:36:17.598022 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:17 crc kubenswrapper[4787]: I1001 09:36:17.598139 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 09:36:17 crc kubenswrapper[4787]: I1001 09:36:17.598968 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:17 crc kubenswrapper[4787]: I1001 09:36:17.599007 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:17 crc kubenswrapper[4787]: I1001 09:36:17.599015 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:17 crc kubenswrapper[4787]: I1001 09:36:17.939675 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 09:36:17 crc kubenswrapper[4787]: I1001 09:36:17.939878 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:17 crc kubenswrapper[4787]: I1001 09:36:17.941010 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:17 crc kubenswrapper[4787]: I1001 09:36:17.941055 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:17 crc kubenswrapper[4787]: I1001 09:36:17.941071 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:18 crc kubenswrapper[4787]: I1001 09:36:18.600511 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:18 crc kubenswrapper[4787]: I1001 09:36:18.601665 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:18 crc kubenswrapper[4787]: I1001 09:36:18.601710 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:18 crc kubenswrapper[4787]: I1001 09:36:18.601723 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:18 crc kubenswrapper[4787]: E1001 09:36:18.611033 4787 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 01 09:36:19 crc kubenswrapper[4787]: I1001 09:36:19.713548 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 09:36:19 crc kubenswrapper[4787]: I1001 09:36:19.713655 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:19 crc kubenswrapper[4787]: I1001 09:36:19.714786 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:19 crc kubenswrapper[4787]: I1001 09:36:19.714852 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:19 crc kubenswrapper[4787]: I1001 09:36:19.714880 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:21 crc kubenswrapper[4787]: I1001 09:36:21.800670 4787 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 01 09:36:21 crc kubenswrapper[4787]: I1001 09:36:21.800756 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 01 09:36:22 crc kubenswrapper[4787]: I1001 09:36:22.455622 4787 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 01 09:36:22 crc kubenswrapper[4787]: W1001 09:36:22.830481 4787 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 01 09:36:22 crc kubenswrapper[4787]: I1001 09:36:22.830565 4787 trace.go:236] Trace[1649987025]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Oct-2025 09:36:12.829) (total time: 10001ms): Oct 01 09:36:22 crc kubenswrapper[4787]: Trace[1649987025]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (09:36:22.830) Oct 01 09:36:22 crc kubenswrapper[4787]: Trace[1649987025]: [10.001202345s] [10.001202345s] END Oct 01 09:36:22 crc kubenswrapper[4787]: E1001 09:36:22.830583 4787 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 01 09:36:22 crc kubenswrapper[4787]: I1001 09:36:22.882737 4787 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 01 09:36:22 crc kubenswrapper[4787]: I1001 09:36:22.882805 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 01 09:36:22 crc kubenswrapper[4787]: I1001 09:36:22.887978 4787 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 01 09:36:22 crc kubenswrapper[4787]: I1001 09:36:22.888043 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 01 09:36:23 crc kubenswrapper[4787]: I1001 09:36:23.953514 4787 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 01 09:36:23 crc kubenswrapper[4787]: [+]log ok Oct 01 09:36:23 crc kubenswrapper[4787]: [+]etcd ok Oct 01 09:36:23 crc kubenswrapper[4787]: [+]poststarthook/openshift.io-startkubeinformers ok Oct 01 09:36:23 crc kubenswrapper[4787]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Oct 01 09:36:23 crc kubenswrapper[4787]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Oct 01 09:36:23 crc kubenswrapper[4787]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 01 09:36:23 crc kubenswrapper[4787]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 01 09:36:23 crc kubenswrapper[4787]: [+]poststarthook/openshift.io-api-request-count-filter ok Oct 01 09:36:23 crc kubenswrapper[4787]: [+]poststarthook/generic-apiserver-start-informers ok Oct 01 09:36:23 crc kubenswrapper[4787]: [+]poststarthook/priority-and-fairness-config-consumer ok Oct 01 09:36:23 crc kubenswrapper[4787]: [+]poststarthook/priority-and-fairness-filter ok Oct 01 09:36:23 crc kubenswrapper[4787]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 01 09:36:23 crc kubenswrapper[4787]: [+]poststarthook/start-apiextensions-informers ok Oct 01 09:36:23 crc kubenswrapper[4787]: [+]poststarthook/start-apiextensions-controllers ok Oct 01 09:36:23 crc kubenswrapper[4787]: [+]poststarthook/crd-informer-synced ok Oct 01 09:36:23 crc kubenswrapper[4787]: [+]poststarthook/start-system-namespaces-controller ok Oct 01 09:36:23 crc kubenswrapper[4787]: [+]poststarthook/start-cluster-authentication-info-controller ok Oct 01 09:36:23 crc kubenswrapper[4787]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Oct 01 09:36:23 crc kubenswrapper[4787]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Oct 01 09:36:23 crc kubenswrapper[4787]: [+]poststarthook/start-legacy-token-tracking-controller ok Oct 01 09:36:23 crc kubenswrapper[4787]: [+]poststarthook/start-service-ip-repair-controllers ok Oct 01 09:36:23 crc kubenswrapper[4787]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Oct 01 09:36:23 crc kubenswrapper[4787]: [+]poststarthook/scheduling/bootstrap-system-priority-classes ok Oct 01 09:36:23 crc kubenswrapper[4787]: [+]poststarthook/priority-and-fairness-config-producer ok Oct 01 09:36:23 crc kubenswrapper[4787]: [+]poststarthook/bootstrap-controller ok Oct 01 09:36:23 crc kubenswrapper[4787]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Oct 01 09:36:23 crc kubenswrapper[4787]: [+]poststarthook/start-kube-aggregator-informers ok Oct 01 09:36:23 crc kubenswrapper[4787]: [+]poststarthook/apiservice-status-local-available-controller ok Oct 01 09:36:23 crc kubenswrapper[4787]: [+]poststarthook/apiservice-status-remote-available-controller ok Oct 01 09:36:23 crc kubenswrapper[4787]: [+]poststarthook/apiservice-registration-controller ok Oct 01 09:36:23 crc kubenswrapper[4787]: [+]poststarthook/apiservice-wait-for-first-sync ok Oct 01 09:36:23 crc kubenswrapper[4787]: [+]poststarthook/apiservice-discovery-controller ok Oct 01 09:36:23 crc kubenswrapper[4787]: [+]poststarthook/kube-apiserver-autoregistration ok Oct 01 09:36:23 crc kubenswrapper[4787]: [+]autoregister-completion ok Oct 01 09:36:23 crc kubenswrapper[4787]: [+]poststarthook/apiservice-openapi-controller ok Oct 01 09:36:23 crc kubenswrapper[4787]: [+]poststarthook/apiservice-openapiv3-controller ok Oct 01 09:36:23 crc kubenswrapper[4787]: livez check failed Oct 01 09:36:23 crc kubenswrapper[4787]: I1001 09:36:23.953607 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 09:36:25 crc kubenswrapper[4787]: I1001 09:36:25.438230 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 01 09:36:25 crc kubenswrapper[4787]: I1001 09:36:25.438406 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:25 crc kubenswrapper[4787]: I1001 09:36:25.439579 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:25 crc kubenswrapper[4787]: I1001 09:36:25.439616 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:25 crc kubenswrapper[4787]: I1001 09:36:25.439630 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:25 crc kubenswrapper[4787]: I1001 09:36:25.454009 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 01 09:36:25 crc kubenswrapper[4787]: I1001 09:36:25.619186 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:25 crc kubenswrapper[4787]: I1001 09:36:25.620058 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:25 crc kubenswrapper[4787]: I1001 09:36:25.620114 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:25 crc kubenswrapper[4787]: I1001 09:36:25.620127 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:26 crc kubenswrapper[4787]: I1001 09:36:26.207295 4787 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 01 09:36:26 crc kubenswrapper[4787]: I1001 09:36:26.207363 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 01 09:36:27 crc kubenswrapper[4787]: E1001 09:36:27.878639 4787 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 01 09:36:27 crc kubenswrapper[4787]: I1001 09:36:27.880788 4787 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 01 09:36:27 crc kubenswrapper[4787]: E1001 09:36:27.882953 4787 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 01 09:36:27 crc kubenswrapper[4787]: I1001 09:36:27.883264 4787 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 01 09:36:27 crc kubenswrapper[4787]: I1001 09:36:27.883417 4787 trace.go:236] Trace[1785374765]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Oct-2025 09:36:15.100) (total time: 12783ms): Oct 01 09:36:27 crc kubenswrapper[4787]: Trace[1785374765]: ---"Objects listed" error: 12782ms (09:36:27.883) Oct 01 09:36:27 crc kubenswrapper[4787]: Trace[1785374765]: [12.783020616s] [12.783020616s] END Oct 01 09:36:27 crc kubenswrapper[4787]: I1001 09:36:27.883438 4787 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 01 09:36:27 crc kubenswrapper[4787]: I1001 09:36:27.884219 4787 trace.go:236] Trace[631107025]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Oct-2025 09:36:16.719) (total time: 11165ms): Oct 01 09:36:27 crc kubenswrapper[4787]: Trace[631107025]: ---"Objects listed" error: 11164ms (09:36:27.883) Oct 01 09:36:27 crc kubenswrapper[4787]: Trace[631107025]: [11.165156388s] [11.165156388s] END Oct 01 09:36:27 crc kubenswrapper[4787]: I1001 09:36:27.884251 4787 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 01 09:36:27 crc kubenswrapper[4787]: I1001 09:36:27.975587 4787 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:45530->192.168.126.11:17697: read: connection reset by peer" start-of-body= Oct 01 09:36:27 crc kubenswrapper[4787]: I1001 09:36:27.975655 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:45530->192.168.126.11:17697: read: connection reset by peer" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.450618 4787 apiserver.go:52] "Watching apiserver" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.453591 4787 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.453845 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-dns/node-resolver-cfzj5","openshift-image-registry/node-ca-5ghjd","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf"] Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.454200 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.454370 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.454390 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:36:28 crc kubenswrapper[4787]: E1001 09:36:28.454416 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:36:28 crc kubenswrapper[4787]: E1001 09:36:28.454461 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.454482 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.454525 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.454634 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:36:28 crc kubenswrapper[4787]: E1001 09:36:28.454688 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.455091 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-cfzj5" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.455177 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-5ghjd" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.456034 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.457126 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.457126 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.457579 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.458177 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.460786 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.460954 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.461354 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.461617 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.461724 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.461738 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.461774 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.463283 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.463474 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.463482 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.463741 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.469850 4787 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.474655 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.475527 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-8vh6d"] Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.476560 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.480052 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.480060 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.480922 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.481107 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.481433 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.482659 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-kzcns"] Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.483274 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-kzcns" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.485412 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-z8tjz"] Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.485871 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.487770 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.487810 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.487844 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.487876 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.487901 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.487929 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.487957 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.491874 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.491912 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.491942 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.491967 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.492106 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.492148 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.492177 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.492206 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.492254 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.492300 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.492369 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.492398 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.492460 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-dmsxm"] Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.488479 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.488563 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.488679 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.489340 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.489984 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.490217 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.490367 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.493447 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.493480 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.493524 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.493568 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.490419 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.490658 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.490894 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.490965 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.491354 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.491763 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.491794 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.492619 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.493605 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.494324 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.494336 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.494371 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.494628 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.494662 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.494962 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.495047 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.496170 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.496393 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.496506 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.497294 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.498053 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.499364 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.499406 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.499430 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.499453 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.499473 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.499493 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.499498 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.499515 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.499534 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.499557 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.499577 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.499597 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.499616 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.499636 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.499657 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.499678 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.499699 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.499713 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.499720 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.499762 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.499787 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.499811 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.499866 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.499878 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.499916 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.499932 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.499942 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.499965 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.499990 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500013 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500036 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500066 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500110 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500124 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500136 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500159 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500184 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500213 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500236 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500281 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500303 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500327 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500349 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500371 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500393 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500415 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500435 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500458 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500481 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500503 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500542 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500564 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500586 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500608 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500629 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500652 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500679 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500701 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500726 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500750 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500795 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500819 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500842 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500865 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500886 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500909 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500947 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500985 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501023 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501046 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501068 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501108 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501150 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501175 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501196 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501220 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501244 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501264 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501286 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501315 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501336 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501359 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501383 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501406 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501431 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501450 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501490 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501511 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501532 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501552 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501572 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501596 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501639 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501663 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501685 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501707 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501729 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501749 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501770 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501792 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501820 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501842 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501865 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501888 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501910 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501931 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501951 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501973 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502013 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502038 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502062 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502104 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502128 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502151 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502138 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502172 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502196 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502228 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502251 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502274 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502295 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502320 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502343 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502365 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502387 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502409 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502433 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502461 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502484 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502505 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502527 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502551 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502572 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502598 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502621 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502646 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502671 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502697 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502794 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502822 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502848 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502874 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502929 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502952 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502980 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.503003 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.503026 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.503046 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.503067 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504071 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504125 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504150 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504174 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504200 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504224 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504248 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504273 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504302 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504329 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504359 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504384 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504408 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504431 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504459 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504486 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504512 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504538 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504563 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504588 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504612 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504635 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504658 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504684 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504707 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504731 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504756 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504781 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504808 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504860 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504888 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504916 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504945 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504972 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mvlr\" (UniqueName: \"kubernetes.io/projected/3bba21c1-f38f-418b-add8-bff4953fe9a3-kube-api-access-9mvlr\") pod \"node-resolver-cfzj5\" (UID: \"3bba21c1-f38f-418b-add8-bff4953fe9a3\") " pod="openshift-dns/node-resolver-cfzj5" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504997 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505020 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3bba21c1-f38f-418b-add8-bff4953fe9a3-hosts-file\") pod \"node-resolver-cfzj5\" (UID: \"3bba21c1-f38f-418b-add8-bff4953fe9a3\") " pod="openshift-dns/node-resolver-cfzj5" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505050 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505089 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/f9b99ece-c0d3-4482-ab68-2672f03b107b-serviceca\") pod \"node-ca-5ghjd\" (UID: \"f9b99ece-c0d3-4482-ab68-2672f03b107b\") " pod="openshift-image-registry/node-ca-5ghjd" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505119 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505144 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505167 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505190 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bphdx\" (UniqueName: \"kubernetes.io/projected/f9b99ece-c0d3-4482-ab68-2672f03b107b-kube-api-access-bphdx\") pod \"node-ca-5ghjd\" (UID: \"f9b99ece-c0d3-4482-ab68-2672f03b107b\") " pod="openshift-image-registry/node-ca-5ghjd" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505212 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505238 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505258 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505284 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505307 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f9b99ece-c0d3-4482-ab68-2672f03b107b-host\") pod \"node-ca-5ghjd\" (UID: \"f9b99ece-c0d3-4482-ab68-2672f03b107b\") " pod="openshift-image-registry/node-ca-5ghjd" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505330 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505373 4787 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505391 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505406 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505421 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505436 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505451 4787 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505465 4787 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505478 4787 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505492 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505507 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505522 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505536 4787 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505549 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505566 4787 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505579 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505594 4787 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505626 4787 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505643 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505655 4787 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505669 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505682 4787 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505695 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505710 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505724 4787 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505736 4787 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.506167 4787 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.514820 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500134 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500236 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500397 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500538 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500775 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.500831 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501104 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501207 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501487 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501592 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.501715 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502155 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502297 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.528723 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502334 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502515 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502564 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502684 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502796 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.502939 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.503049 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.503760 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.503867 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504023 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504068 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504066 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504281 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504519 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504646 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.528977 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504779 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.504855 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505014 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505247 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505275 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.505501 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.508545 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.508960 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.508981 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.509272 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.509388 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.509692 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.510000 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.510313 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.510322 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.510394 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.510501 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.510592 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.510848 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.510986 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.511275 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.512137 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.514199 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.514647 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.514877 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.515173 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.515358 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.516274 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.517212 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.517502 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.517631 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.517648 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.517667 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.517969 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.518138 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.518344 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.518371 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.518447 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.518648 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.518645 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.519001 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.519370 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.519448 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.519593 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.519906 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.520471 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.521836 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.522605 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.529111 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.529183 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.529435 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.529466 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.529518 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.529583 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.529621 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.529637 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.529726 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.530026 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.532891 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.532996 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.533019 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.530743 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.530058 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.529760 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.530227 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.530358 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.530593 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.531099 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.531431 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.531473 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.533600 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.533836 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.534033 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.534317 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.534555 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.534838 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.535049 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.535361 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.535622 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.535968 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.536249 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.536748 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.538608 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.538861 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.539205 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.539358 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.539513 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: E1001 09:36:28.539525 4787 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.539529 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.541227 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.539868 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.539891 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.540236 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.540418 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.540501 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.540794 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.540846 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.540989 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.541142 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.541292 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.540329 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: E1001 09:36:28.541474 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 09:36:29.041433636 +0000 UTC m=+21.156577883 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.541524 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.541676 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.542116 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.542407 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.542735 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.543282 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.543688 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.543704 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.543877 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.544009 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.544312 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.544868 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.545567 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.546367 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.546630 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.546893 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.547107 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.547487 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.547614 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.547701 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.547861 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.547980 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.548176 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.548181 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.548291 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: E1001 09:36:28.548375 4787 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.548428 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: E1001 09:36:28.548545 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 09:36:29.048526664 +0000 UTC m=+21.163670821 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 09:36:28 crc kubenswrapper[4787]: E1001 09:36:28.548842 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:36:29.048821336 +0000 UTC m=+21.163965583 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.548895 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.549060 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.550753 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: E1001 09:36:28.551521 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 09:36:28 crc kubenswrapper[4787]: E1001 09:36:28.554247 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 09:36:28 crc kubenswrapper[4787]: E1001 09:36:28.554274 4787 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.552562 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: E1001 09:36:28.554319 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-01 09:36:29.054307641 +0000 UTC m=+21.169451798 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.552620 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.555315 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.557349 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.557569 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.558962 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.557869 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: E1001 09:36:28.560365 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 09:36:28 crc kubenswrapper[4787]: E1001 09:36:28.560392 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 09:36:28 crc kubenswrapper[4787]: E1001 09:36:28.560406 4787 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:36:28 crc kubenswrapper[4787]: E1001 09:36:28.560457 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-01 09:36:29.060436851 +0000 UTC m=+21.175581108 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.560691 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.560925 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.561007 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.561195 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.561900 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.564437 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.565037 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.565311 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.565787 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.565894 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.566141 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.566419 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.568818 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.569773 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.570029 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.571737 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.575833 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.576561 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.576571 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.576832 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.577620 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.577668 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.578326 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.579630 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.579964 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.580257 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.580598 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.581157 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.582753 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.585944 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.586756 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.589015 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.591929 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.597939 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.599023 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.601413 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.602264 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.603453 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.603789 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.607000 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.607551 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-host-run-k8s-cni-cncf-io\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.607600 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/eadf2545-886e-4642-b909-704c6cd4134e-env-overrides\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.607682 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3bba21c1-f38f-418b-add8-bff4953fe9a3-hosts-file\") pod \"node-resolver-cfzj5\" (UID: \"3bba21c1-f38f-418b-add8-bff4953fe9a3\") " pod="openshift-dns/node-resolver-cfzj5" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.607707 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-host-var-lib-cni-bin\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.607726 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-host-run-multus-certs\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.607745 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-etc-kubernetes\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.607760 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-run-ovn-kubernetes\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.607778 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-multus-socket-dir-parent\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.607833 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/f9b99ece-c0d3-4482-ab68-2672f03b107b-serviceca\") pod \"node-ca-5ghjd\" (UID: \"f9b99ece-c0d3-4482-ab68-2672f03b107b\") " pod="openshift-image-registry/node-ca-5ghjd" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.609519 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3bba21c1-f38f-418b-add8-bff4953fe9a3-hosts-file\") pod \"node-resolver-cfzj5\" (UID: \"3bba21c1-f38f-418b-add8-bff4953fe9a3\") " pod="openshift-dns/node-resolver-cfzj5" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.609591 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wcmv\" (UniqueName: \"kubernetes.io/projected/53f5ffd2-7649-4c9e-929f-a68187409644-kube-api-access-2wcmv\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.610025 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/60771b6a-9398-4987-bb28-a20ceb6cca38-tuning-conf-dir\") pod \"multus-additional-cni-plugins-kzcns\" (UID: \"60771b6a-9398-4987-bb28-a20ceb6cca38\") " pod="openshift-multus/multus-additional-cni-plugins-kzcns" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.610132 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9d7sz\" (UniqueName: \"kubernetes.io/projected/60771b6a-9398-4987-bb28-a20ceb6cca38-kube-api-access-9d7sz\") pod \"multus-additional-cni-plugins-kzcns\" (UID: \"60771b6a-9398-4987-bb28-a20ceb6cca38\") " pod="openshift-multus/multus-additional-cni-plugins-kzcns" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.610350 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.610416 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-os-release\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.610517 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.610578 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.610630 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.610706 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-multus-cni-dir\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.610781 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.610814 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/60771b6a-9398-4987-bb28-a20ceb6cca38-os-release\") pod \"multus-additional-cni-plugins-kzcns\" (UID: \"60771b6a-9398-4987-bb28-a20ceb6cca38\") " pod="openshift-multus/multus-additional-cni-plugins-kzcns" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.610968 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-cnibin\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.610985 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/f9b99ece-c0d3-4482-ab68-2672f03b107b-serviceca\") pod \"node-ca-5ghjd\" (UID: \"f9b99ece-c0d3-4482-ab68-2672f03b107b\") " pod="openshift-image-registry/node-ca-5ghjd" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.610990 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-host-var-lib-cni-multus\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.611136 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-hostroot\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.611214 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-slash\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.611249 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-cni-netd\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.611277 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-systemd-units\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.611314 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-system-cni-dir\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.611340 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-host-var-lib-kubelet\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.611372 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/fa51b883-5f3e-4141-9d4c-37704aac7718-rootfs\") pod \"machine-config-daemon-8vh6d\" (UID: \"fa51b883-5f3e-4141-9d4c-37704aac7718\") " pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.611401 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fa51b883-5f3e-4141-9d4c-37704aac7718-proxy-tls\") pod \"machine-config-daemon-8vh6d\" (UID: \"fa51b883-5f3e-4141-9d4c-37704aac7718\") " pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.611420 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-kubelet\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.611447 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-run-systemd\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.611487 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-node-log\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.611509 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/60771b6a-9398-4987-bb28-a20ceb6cca38-cnibin\") pod \"multus-additional-cni-plugins-kzcns\" (UID: \"60771b6a-9398-4987-bb28-a20ceb6cca38\") " pod="openshift-multus/multus-additional-cni-plugins-kzcns" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.611535 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppfbw\" (UniqueName: \"kubernetes.io/projected/fa51b883-5f3e-4141-9d4c-37704aac7718-kube-api-access-ppfbw\") pod \"machine-config-daemon-8vh6d\" (UID: \"fa51b883-5f3e-4141-9d4c-37704aac7718\") " pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.611557 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-log-socket\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.611574 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/eadf2545-886e-4642-b909-704c6cd4134e-ovnkube-config\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.611593 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/eadf2545-886e-4642-b909-704c6cd4134e-ovn-node-metrics-cert\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.611615 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/eadf2545-886e-4642-b909-704c6cd4134e-ovnkube-script-lib\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.611682 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fa51b883-5f3e-4141-9d4c-37704aac7718-mcd-auth-proxy-config\") pod \"machine-config-daemon-8vh6d\" (UID: \"fa51b883-5f3e-4141-9d4c-37704aac7718\") " pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.611705 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bphdx\" (UniqueName: \"kubernetes.io/projected/f9b99ece-c0d3-4482-ab68-2672f03b107b-kube-api-access-bphdx\") pod \"node-ca-5ghjd\" (UID: \"f9b99ece-c0d3-4482-ab68-2672f03b107b\") " pod="openshift-image-registry/node-ca-5ghjd" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.611728 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/53f5ffd2-7649-4c9e-929f-a68187409644-cni-binary-copy\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.611748 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/53f5ffd2-7649-4c9e-929f-a68187409644-multus-daemon-config\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.611765 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-run-ovn\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.611783 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/60771b6a-9398-4987-bb28-a20ceb6cca38-cni-binary-copy\") pod \"multus-additional-cni-plugins-kzcns\" (UID: \"60771b6a-9398-4987-bb28-a20ceb6cca38\") " pod="openshift-multus/multus-additional-cni-plugins-kzcns" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.611800 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/60771b6a-9398-4987-bb28-a20ceb6cca38-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-kzcns\" (UID: \"60771b6a-9398-4987-bb28-a20ceb6cca38\") " pod="openshift-multus/multus-additional-cni-plugins-kzcns" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.611854 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f9b99ece-c0d3-4482-ab68-2672f03b107b-host\") pod \"node-ca-5ghjd\" (UID: \"f9b99ece-c0d3-4482-ab68-2672f03b107b\") " pod="openshift-image-registry/node-ca-5ghjd" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.611896 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-run-netns\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.611917 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-cni-bin\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.611944 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-multus-conf-dir\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.611960 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-var-lib-openvswitch\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.612006 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-etc-openvswitch\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.612023 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-run-openvswitch\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.612601 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/60771b6a-9398-4987-bb28-a20ceb6cca38-system-cni-dir\") pod \"multus-additional-cni-plugins-kzcns\" (UID: \"60771b6a-9398-4987-bb28-a20ceb6cca38\") " pod="openshift-multus/multus-additional-cni-plugins-kzcns" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.612651 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8kgg\" (UniqueName: \"kubernetes.io/projected/eadf2545-886e-4642-b909-704c6cd4134e-kube-api-access-z8kgg\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.612682 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mvlr\" (UniqueName: \"kubernetes.io/projected/3bba21c1-f38f-418b-add8-bff4953fe9a3-kube-api-access-9mvlr\") pod \"node-resolver-cfzj5\" (UID: \"3bba21c1-f38f-418b-add8-bff4953fe9a3\") " pod="openshift-dns/node-resolver-cfzj5" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.612714 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-host-run-netns\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.612022 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f9b99ece-c0d3-4482-ab68-2672f03b107b-host\") pod \"node-ca-5ghjd\" (UID: \"f9b99ece-c0d3-4482-ab68-2672f03b107b\") " pod="openshift-image-registry/node-ca-5ghjd" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613314 4787 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613339 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613353 4787 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613364 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613375 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613389 4787 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613399 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613409 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613421 4787 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613431 4787 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613444 4787 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613454 4787 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613464 4787 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613474 4787 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613487 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613498 4787 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613508 4787 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613517 4787 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613527 4787 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613537 4787 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613548 4787 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613558 4787 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613569 4787 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613578 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613588 4787 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613598 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613606 4787 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613616 4787 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613632 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613641 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613651 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613661 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613669 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613680 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613690 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613701 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613714 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613724 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613765 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613776 4787 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613785 4787 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613796 4787 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613806 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613820 4787 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613830 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613840 4787 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613850 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613860 4787 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613869 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613878 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613887 4787 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613899 4787 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613909 4787 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613924 4787 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613933 4787 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613943 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613956 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613967 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613977 4787 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613987 4787 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.613998 4787 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.614008 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.614018 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.614028 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.614037 4787 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.614048 4787 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.614057 4787 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.614065 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.614089 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.614100 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.614110 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.614120 4787 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.614133 4787 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.614142 4787 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.614152 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.614162 4787 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.614172 4787 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.614183 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.614194 4787 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.614204 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.614214 4787 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.614224 4787 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.614234 4787 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.614244 4787 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.614253 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.614263 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.614272 4787 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.614282 4787 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.614296 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.614306 4787 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.614316 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.614327 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.614311 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.614664 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.614339 4787 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.615138 4787 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.615354 4787 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.615602 4787 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.615612 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.615621 4787 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.615630 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.615639 4787 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.615650 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.615659 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.615669 4787 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.615678 4787 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.615686 4787 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.615695 4787 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.615704 4787 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.615713 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.615722 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.615731 4787 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.615740 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.615751 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.615760 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.615770 4787 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.616321 4787 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.616333 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.616344 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.616353 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.616372 4787 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.616382 4787 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.616396 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.616407 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.616417 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.616430 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.616441 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.616452 4787 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.616461 4787 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.616470 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.616479 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.616488 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.616497 4787 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.616510 4787 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.616521 4787 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.616531 4787 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.616540 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.616548 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.616558 4787 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.616568 4787 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.616576 4787 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.616586 4787 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.616595 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.616604 4787 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.616614 4787 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.617196 4787 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.617241 4787 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.617259 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.617278 4787 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.617297 4787 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.617315 4787 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.617335 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.617351 4787 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.617368 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.617381 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.617394 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.617408 4787 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.617422 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.617436 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.617449 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.617463 4787 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.617478 4787 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.617490 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.617503 4787 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.617516 4787 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.617528 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.617540 4787 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.617552 4787 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.617565 4787 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.617579 4787 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.618356 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.622910 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.624403 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.625163 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.627517 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bphdx\" (UniqueName: \"kubernetes.io/projected/f9b99ece-c0d3-4482-ab68-2672f03b107b-kube-api-access-bphdx\") pod \"node-ca-5ghjd\" (UID: \"f9b99ece-c0d3-4482-ab68-2672f03b107b\") " pod="openshift-image-registry/node-ca-5ghjd" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.627771 4787 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.629406 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.631439 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.633770 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.634606 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.635232 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.636205 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.636377 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mvlr\" (UniqueName: \"kubernetes.io/projected/3bba21c1-f38f-418b-add8-bff4953fe9a3-kube-api-access-9mvlr\") pod \"node-resolver-cfzj5\" (UID: \"3bba21c1-f38f-418b-add8-bff4953fe9a3\") " pod="openshift-dns/node-resolver-cfzj5" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.636781 4787 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c" exitCode=255 Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.638170 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.639601 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.640653 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.641516 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.641778 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.642803 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.643575 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.644853 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.645915 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.647789 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.648330 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.649050 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.650699 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.651601 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.651678 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.654789 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.655428 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.656350 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.657063 4787 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.657177 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.657764 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.658295 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.659340 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c"} Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.660907 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.671378 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.684654 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.693728 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.704829 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.714240 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.718856 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-multus-socket-dir-parent\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.718898 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-host-var-lib-cni-bin\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.718918 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-host-run-multus-certs\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.718936 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-etc-kubernetes\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.718951 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-run-ovn-kubernetes\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.718973 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wcmv\" (UniqueName: \"kubernetes.io/projected/53f5ffd2-7649-4c9e-929f-a68187409644-kube-api-access-2wcmv\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.718989 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/60771b6a-9398-4987-bb28-a20ceb6cca38-tuning-conf-dir\") pod \"multus-additional-cni-plugins-kzcns\" (UID: \"60771b6a-9398-4987-bb28-a20ceb6cca38\") " pod="openshift-multus/multus-additional-cni-plugins-kzcns" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.719006 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-os-release\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.719002 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-multus-socket-dir-parent\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.719024 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9d7sz\" (UniqueName: \"kubernetes.io/projected/60771b6a-9398-4987-bb28-a20ceb6cca38-kube-api-access-9d7sz\") pod \"multus-additional-cni-plugins-kzcns\" (UID: \"60771b6a-9398-4987-bb28-a20ceb6cca38\") " pod="openshift-multus/multus-additional-cni-plugins-kzcns" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.719156 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-multus-cni-dir\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.719180 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.719202 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/60771b6a-9398-4987-bb28-a20ceb6cca38-os-release\") pod \"multus-additional-cni-plugins-kzcns\" (UID: \"60771b6a-9398-4987-bb28-a20ceb6cca38\") " pod="openshift-multus/multus-additional-cni-plugins-kzcns" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.719240 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-cnibin\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.719260 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-host-var-lib-cni-multus\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.719280 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-hostroot\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.719304 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-systemd-units\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.719313 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-host-var-lib-cni-bin\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.719355 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-slash\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.719387 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-host-run-multus-certs\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.719410 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-etc-kubernetes\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.719435 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-run-ovn-kubernetes\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.719483 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-multus-cni-dir\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.719517 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.719319 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-slash\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.719702 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/60771b6a-9398-4987-bb28-a20ceb6cca38-os-release\") pod \"multus-additional-cni-plugins-kzcns\" (UID: \"60771b6a-9398-4987-bb28-a20ceb6cca38\") " pod="openshift-multus/multus-additional-cni-plugins-kzcns" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.719725 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-os-release\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.719785 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-cni-netd\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.719843 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-cni-netd\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.719851 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-cnibin\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.719843 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-systemd-units\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.719869 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-host-var-lib-cni-multus\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.719880 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/fa51b883-5f3e-4141-9d4c-37704aac7718-rootfs\") pod \"machine-config-daemon-8vh6d\" (UID: \"fa51b883-5f3e-4141-9d4c-37704aac7718\") " pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.719919 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-hostroot\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.719909 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-system-cni-dir\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.719961 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-host-var-lib-kubelet\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.719997 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-host-var-lib-kubelet\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.720027 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/fa51b883-5f3e-4141-9d4c-37704aac7718-rootfs\") pod \"machine-config-daemon-8vh6d\" (UID: \"fa51b883-5f3e-4141-9d4c-37704aac7718\") " pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.720031 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fa51b883-5f3e-4141-9d4c-37704aac7718-proxy-tls\") pod \"machine-config-daemon-8vh6d\" (UID: \"fa51b883-5f3e-4141-9d4c-37704aac7718\") " pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.720058 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-kubelet\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.720117 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/60771b6a-9398-4987-bb28-a20ceb6cca38-tuning-conf-dir\") pod \"multus-additional-cni-plugins-kzcns\" (UID: \"60771b6a-9398-4987-bb28-a20ceb6cca38\") " pod="openshift-multus/multus-additional-cni-plugins-kzcns" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.720127 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-kubelet\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.720145 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-run-systemd\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.720186 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-system-cni-dir\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.720182 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-node-log\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.720224 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/60771b6a-9398-4987-bb28-a20ceb6cca38-cnibin\") pod \"multus-additional-cni-plugins-kzcns\" (UID: \"60771b6a-9398-4987-bb28-a20ceb6cca38\") " pod="openshift-multus/multus-additional-cni-plugins-kzcns" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.720161 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-run-systemd\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.720245 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fa51b883-5f3e-4141-9d4c-37704aac7718-mcd-auth-proxy-config\") pod \"machine-config-daemon-8vh6d\" (UID: \"fa51b883-5f3e-4141-9d4c-37704aac7718\") " pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.720264 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppfbw\" (UniqueName: \"kubernetes.io/projected/fa51b883-5f3e-4141-9d4c-37704aac7718-kube-api-access-ppfbw\") pod \"machine-config-daemon-8vh6d\" (UID: \"fa51b883-5f3e-4141-9d4c-37704aac7718\") " pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.720270 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/60771b6a-9398-4987-bb28-a20ceb6cca38-cnibin\") pod \"multus-additional-cni-plugins-kzcns\" (UID: \"60771b6a-9398-4987-bb28-a20ceb6cca38\") " pod="openshift-multus/multus-additional-cni-plugins-kzcns" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.720282 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-log-socket\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.720300 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/eadf2545-886e-4642-b909-704c6cd4134e-ovnkube-config\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.720316 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/eadf2545-886e-4642-b909-704c6cd4134e-ovn-node-metrics-cert\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.720338 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/eadf2545-886e-4642-b909-704c6cd4134e-ovnkube-script-lib\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.720367 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/53f5ffd2-7649-4c9e-929f-a68187409644-cni-binary-copy\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.720384 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/53f5ffd2-7649-4c9e-929f-a68187409644-multus-daemon-config\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.720399 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-run-ovn\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.720414 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/60771b6a-9398-4987-bb28-a20ceb6cca38-cni-binary-copy\") pod \"multus-additional-cni-plugins-kzcns\" (UID: \"60771b6a-9398-4987-bb28-a20ceb6cca38\") " pod="openshift-multus/multus-additional-cni-plugins-kzcns" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.720433 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/60771b6a-9398-4987-bb28-a20ceb6cca38-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-kzcns\" (UID: \"60771b6a-9398-4987-bb28-a20ceb6cca38\") " pod="openshift-multus/multus-additional-cni-plugins-kzcns" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.720456 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-run-netns\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.720471 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-cni-bin\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.720496 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-multus-conf-dir\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.720511 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-var-lib-openvswitch\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.720528 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-etc-openvswitch\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.720545 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-run-openvswitch\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.720559 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/60771b6a-9398-4987-bb28-a20ceb6cca38-system-cni-dir\") pod \"multus-additional-cni-plugins-kzcns\" (UID: \"60771b6a-9398-4987-bb28-a20ceb6cca38\") " pod="openshift-multus/multus-additional-cni-plugins-kzcns" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.720574 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-host-run-netns\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.720591 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8kgg\" (UniqueName: \"kubernetes.io/projected/eadf2545-886e-4642-b909-704c6cd4134e-kube-api-access-z8kgg\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.720608 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-host-run-k8s-cni-cncf-io\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.720623 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/eadf2545-886e-4642-b909-704c6cd4134e-env-overrides\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.720668 4787 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.720872 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fa51b883-5f3e-4141-9d4c-37704aac7718-mcd-auth-proxy-config\") pod \"machine-config-daemon-8vh6d\" (UID: \"fa51b883-5f3e-4141-9d4c-37704aac7718\") " pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.720927 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-etc-openvswitch\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.720963 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-run-netns\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.720995 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-cni-bin\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.721030 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-multus-conf-dir\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.721064 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-var-lib-openvswitch\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.721121 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-log-socket\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.721299 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/eadf2545-886e-4642-b909-704c6cd4134e-env-overrides\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.721355 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-run-openvswitch\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.721380 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/60771b6a-9398-4987-bb28-a20ceb6cca38-system-cni-dir\") pod \"multus-additional-cni-plugins-kzcns\" (UID: \"60771b6a-9398-4987-bb28-a20ceb6cca38\") " pod="openshift-multus/multus-additional-cni-plugins-kzcns" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.721415 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-host-run-netns\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.721647 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/eadf2545-886e-4642-b909-704c6cd4134e-ovnkube-script-lib\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.721740 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/60771b6a-9398-4987-bb28-a20ceb6cca38-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-kzcns\" (UID: \"60771b6a-9398-4987-bb28-a20ceb6cca38\") " pod="openshift-multus/multus-additional-cni-plugins-kzcns" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.721790 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-run-ovn\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.722148 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/53f5ffd2-7649-4c9e-929f-a68187409644-multus-daemon-config\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.722212 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-node-log\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.722222 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/53f5ffd2-7649-4c9e-929f-a68187409644-cni-binary-copy\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.722265 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/53f5ffd2-7649-4c9e-929f-a68187409644-host-run-k8s-cni-cncf-io\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.722435 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/60771b6a-9398-4987-bb28-a20ceb6cca38-cni-binary-copy\") pod \"multus-additional-cni-plugins-kzcns\" (UID: \"60771b6a-9398-4987-bb28-a20ceb6cca38\") " pod="openshift-multus/multus-additional-cni-plugins-kzcns" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.722491 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/eadf2545-886e-4642-b909-704c6cd4134e-ovnkube-config\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.725261 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fa51b883-5f3e-4141-9d4c-37704aac7718-proxy-tls\") pod \"machine-config-daemon-8vh6d\" (UID: \"fa51b883-5f3e-4141-9d4c-37704aac7718\") " pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.726317 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.729588 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/eadf2545-886e-4642-b909-704c6cd4134e-ovn-node-metrics-cert\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.740628 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wcmv\" (UniqueName: \"kubernetes.io/projected/53f5ffd2-7649-4c9e-929f-a68187409644-kube-api-access-2wcmv\") pod \"multus-z8tjz\" (UID: \"53f5ffd2-7649-4c9e-929f-a68187409644\") " pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.741159 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppfbw\" (UniqueName: \"kubernetes.io/projected/fa51b883-5f3e-4141-9d4c-37704aac7718-kube-api-access-ppfbw\") pod \"machine-config-daemon-8vh6d\" (UID: \"fa51b883-5f3e-4141-9d4c-37704aac7718\") " pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.744634 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8kgg\" (UniqueName: \"kubernetes.io/projected/eadf2545-886e-4642-b909-704c6cd4134e-kube-api-access-z8kgg\") pod \"ovnkube-node-dmsxm\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.744672 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9d7sz\" (UniqueName: \"kubernetes.io/projected/60771b6a-9398-4987-bb28-a20ceb6cca38-kube-api-access-9d7sz\") pod \"multus-additional-cni-plugins-kzcns\" (UID: \"60771b6a-9398-4987-bb28-a20ceb6cca38\") " pod="openshift-multus/multus-additional-cni-plugins-kzcns" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.748989 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.757409 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.765790 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.766825 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.773823 4787 scope.go:117] "RemoveContainer" containerID="149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.774139 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.779317 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:28 crc kubenswrapper[4787]: W1001 09:36:28.779474 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-0e1df38aef29f1bb502b3e448a5413823082a85c56ad9697c524dc1283a23934 WatchSource:0}: Error finding container 0e1df38aef29f1bb502b3e448a5413823082a85c56ad9697c524dc1283a23934: Status 404 returned error can't find the container with id 0e1df38aef29f1bb502b3e448a5413823082a85c56ad9697c524dc1283a23934 Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.784136 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.789694 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.794166 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.803810 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-cfzj5" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.821136 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.838443 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-5ghjd" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.853392 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.857384 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.876385 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.884635 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-kzcns" Oct 01 09:36:28 crc kubenswrapper[4787]: W1001 09:36:28.886944 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-f4d7c8de5889e9e89f6664cbad8421cb9b6155b8dc02859990ed9d404b73b32e WatchSource:0}: Error finding container f4d7c8de5889e9e89f6664cbad8421cb9b6155b8dc02859990ed9d404b73b32e: Status 404 returned error can't find the container with id f4d7c8de5889e9e89f6664cbad8421cb9b6155b8dc02859990ed9d404b73b32e Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.900035 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.903642 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-z8tjz" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.916447 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.921166 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:28 crc kubenswrapper[4787]: W1001 09:36:28.925212 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf9b99ece_c0d3_4482_ab68_2672f03b107b.slice/crio-1c68d9fb4a1a31de9bafcf87d1f680fc47167bc7e5d12aae9cebf9857d71cd7e WatchSource:0}: Error finding container 1c68d9fb4a1a31de9bafcf87d1f680fc47167bc7e5d12aae9cebf9857d71cd7e: Status 404 returned error can't find the container with id 1c68d9fb4a1a31de9bafcf87d1f680fc47167bc7e5d12aae9cebf9857d71cd7e Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.926018 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:28 crc kubenswrapper[4787]: W1001 09:36:28.933796 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa51b883_5f3e_4141_9d4c_37704aac7718.slice/crio-11824a2ce7562c73b9208d904b47844417663126bb9ee0cbe2ed59416a176e09 WatchSource:0}: Error finding container 11824a2ce7562c73b9208d904b47844417663126bb9ee0cbe2ed59416a176e09: Status 404 returned error can't find the container with id 11824a2ce7562c73b9208d904b47844417663126bb9ee0cbe2ed59416a176e09 Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.936306 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.946262 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.951958 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 09:36:28 crc kubenswrapper[4787]: W1001 09:36:28.957190 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60771b6a_9398_4987_bb28_a20ceb6cca38.slice/crio-5474c0ee33dc0b4d19e4fbaddba72f7f6d67ff2bb75c919f5ceafbbd48288a76 WatchSource:0}: Error finding container 5474c0ee33dc0b4d19e4fbaddba72f7f6d67ff2bb75c919f5ceafbbd48288a76: Status 404 returned error can't find the container with id 5474c0ee33dc0b4d19e4fbaddba72f7f6d67ff2bb75c919f5ceafbbd48288a76 Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.962371 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:28 crc kubenswrapper[4787]: W1001 09:36:28.978919 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeadf2545_886e_4642_b909_704c6cd4134e.slice/crio-2f1d4c83eb343b113400d9faa34fc103f326c7ff17e9312a4aa55d85f97106eb WatchSource:0}: Error finding container 2f1d4c83eb343b113400d9faa34fc103f326c7ff17e9312a4aa55d85f97106eb: Status 404 returned error can't find the container with id 2f1d4c83eb343b113400d9faa34fc103f326c7ff17e9312a4aa55d85f97106eb Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.983889 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:28 crc kubenswrapper[4787]: I1001 09:36:28.998063 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0814bed7-1117-4dff-97ef-a3cd1240e615\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:36:27Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1001 09:36:22.478145 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 09:36:22.478910 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3375976971/tls.crt::/tmp/serving-cert-3375976971/tls.key\\\\\\\"\\\\nI1001 09:36:27.943129 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 09:36:27.955126 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 09:36:27.955153 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 09:36:27.955179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 09:36:27.955185 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 09:36:27.963955 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1001 09:36:27.963979 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1001 09:36:27.963986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964008 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 09:36:27.964015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 09:36:27.964018 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 09:36:27.964022 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1001 09:36:27.966112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.005651 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.022726 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.035339 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.047575 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.061821 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.070652 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.083887 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.119320 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.124536 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.124659 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.124700 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.124729 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.124758 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:36:29 crc kubenswrapper[4787]: E1001 09:36:29.124837 4787 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 09:36:29 crc kubenswrapper[4787]: E1001 09:36:29.124852 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 09:36:29 crc kubenswrapper[4787]: E1001 09:36:29.124872 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 09:36:29 crc kubenswrapper[4787]: E1001 09:36:29.124880 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 09:36:30.124864719 +0000 UTC m=+22.240008876 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 09:36:29 crc kubenswrapper[4787]: E1001 09:36:29.124884 4787 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:36:29 crc kubenswrapper[4787]: E1001 09:36:29.124915 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-01 09:36:30.124905979 +0000 UTC m=+22.240050136 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:36:29 crc kubenswrapper[4787]: E1001 09:36:29.124959 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 09:36:29 crc kubenswrapper[4787]: E1001 09:36:29.124967 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 09:36:29 crc kubenswrapper[4787]: E1001 09:36:29.124975 4787 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:36:29 crc kubenswrapper[4787]: E1001 09:36:29.125023 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-01 09:36:30.12501704 +0000 UTC m=+22.240161197 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:36:29 crc kubenswrapper[4787]: E1001 09:36:29.125051 4787 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 09:36:29 crc kubenswrapper[4787]: E1001 09:36:29.125110 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 09:36:30.125064201 +0000 UTC m=+22.240208358 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 09:36:29 crc kubenswrapper[4787]: E1001 09:36:29.125356 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:36:30.125339113 +0000 UTC m=+22.240483490 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.159228 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.198413 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.241096 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.280273 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.329680 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.371015 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.402596 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0814bed7-1117-4dff-97ef-a3cd1240e615\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:36:27Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1001 09:36:22.478145 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 09:36:22.478910 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3375976971/tls.crt::/tmp/serving-cert-3375976971/tls.key\\\\\\\"\\\\nI1001 09:36:27.943129 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 09:36:27.955126 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 09:36:27.955153 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 09:36:27.955179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 09:36:27.955185 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 09:36:27.963955 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1001 09:36:27.963979 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1001 09:36:27.963986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964008 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 09:36:27.964015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 09:36:27.964018 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 09:36:27.964022 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1001 09:36:27.966112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.440449 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.481935 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.520282 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.523484 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:36:29 crc kubenswrapper[4787]: E1001 09:36:29.523643 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.560394 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.640987 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" event={"ID":"fa51b883-5f3e-4141-9d4c-37704aac7718","Type":"ContainerStarted","Data":"0d95f8b1aac879d1746d8a8af7e1d5e672b36c31d5b95c23fee70b237bc3716e"} Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.641055 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" event={"ID":"fa51b883-5f3e-4141-9d4c-37704aac7718","Type":"ContainerStarted","Data":"50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188"} Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.641084 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" event={"ID":"fa51b883-5f3e-4141-9d4c-37704aac7718","Type":"ContainerStarted","Data":"11824a2ce7562c73b9208d904b47844417663126bb9ee0cbe2ed59416a176e09"} Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.642040 4787 generic.go:334] "Generic (PLEG): container finished" podID="eadf2545-886e-4642-b909-704c6cd4134e" containerID="a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c" exitCode=0 Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.642095 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" event={"ID":"eadf2545-886e-4642-b909-704c6cd4134e","Type":"ContainerDied","Data":"a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c"} Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.642123 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" event={"ID":"eadf2545-886e-4642-b909-704c6cd4134e","Type":"ContainerStarted","Data":"2f1d4c83eb343b113400d9faa34fc103f326c7ff17e9312a4aa55d85f97106eb"} Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.643407 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"6bad47df15d6c42d98cec2f1bf55a34e3a599d9fc45ccd6387cd61aa0b26dd93"} Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.644525 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-5ghjd" event={"ID":"f9b99ece-c0d3-4482-ab68-2672f03b107b","Type":"ContainerStarted","Data":"ebc8a6e2b9c6104b0f773e99d68a62e1b2bf0adc05e6f291cf91e7a8cf1bc121"} Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.644552 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-5ghjd" event={"ID":"f9b99ece-c0d3-4482-ab68-2672f03b107b","Type":"ContainerStarted","Data":"1c68d9fb4a1a31de9bafcf87d1f680fc47167bc7e5d12aae9cebf9857d71cd7e"} Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.647985 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.649497 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7c91542e7254c8289b734b3ffa8e8226f2445c915c264e09c179b99604abad1a"} Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.649757 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.651009 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"72701c6061105b7bb81545b5f0b035b220b22b2f8009e82809a037870c625685"} Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.651050 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"0e1df38aef29f1bb502b3e448a5413823082a85c56ad9697c524dc1283a23934"} Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.652700 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-z8tjz" event={"ID":"53f5ffd2-7649-4c9e-929f-a68187409644","Type":"ContainerStarted","Data":"cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885"} Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.652726 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-z8tjz" event={"ID":"53f5ffd2-7649-4c9e-929f-a68187409644","Type":"ContainerStarted","Data":"deaf1a91a01e3e797bdd68010b3bc5182586ed135eb867bd21a60a0a9df03f18"} Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.654352 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-cfzj5" event={"ID":"3bba21c1-f38f-418b-add8-bff4953fe9a3","Type":"ContainerStarted","Data":"a7aa5c938897ea621fc7c71c86e6a1dd2acca273e13afe71e9a27fc81f34cf2e"} Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.654387 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-cfzj5" event={"ID":"3bba21c1-f38f-418b-add8-bff4953fe9a3","Type":"ContainerStarted","Data":"58f72b8dee391d5b8bb7b6fa0e43c06441419e4cc1b5a597e8b24e5b508b6faa"} Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.654504 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.655700 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.656040 4787 generic.go:334] "Generic (PLEG): container finished" podID="60771b6a-9398-4987-bb28-a20ceb6cca38" containerID="785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878" exitCode=0 Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.656118 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" event={"ID":"60771b6a-9398-4987-bb28-a20ceb6cca38","Type":"ContainerDied","Data":"785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878"} Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.656184 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" event={"ID":"60771b6a-9398-4987-bb28-a20ceb6cca38","Type":"ContainerStarted","Data":"5474c0ee33dc0b4d19e4fbaddba72f7f6d67ff2bb75c919f5ceafbbd48288a76"} Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.657712 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"b093e96a244fdec4c5cdce62650d4da31c2ec46207a9c08bbd0a28bf286abb52"} Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.657750 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"4664cb1dde25ada88989e81f0319aedc6dbd363f608a50623efaa3eeb8d00763"} Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.657761 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"f4d7c8de5889e9e89f6664cbad8421cb9b6155b8dc02859990ed9d404b73b32e"} Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.674799 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.689475 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.719584 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.762044 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:29Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.800164 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d95f8b1aac879d1746d8a8af7e1d5e672b36c31d5b95c23fee70b237bc3716e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:29Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.849433 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:29Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.888058 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:29Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.923875 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0814bed7-1117-4dff-97ef-a3cd1240e615\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:36:27Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1001 09:36:22.478145 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 09:36:22.478910 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3375976971/tls.crt::/tmp/serving-cert-3375976971/tls.key\\\\\\\"\\\\nI1001 09:36:27.943129 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 09:36:27.955126 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 09:36:27.955153 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 09:36:27.955179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 09:36:27.955185 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 09:36:27.963955 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1001 09:36:27.963979 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1001 09:36:27.963986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964008 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 09:36:27.964015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 09:36:27.964018 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 09:36:27.964022 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1001 09:36:27.966112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:29Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:29 crc kubenswrapper[4787]: I1001 09:36:29.971636 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:29Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.003426 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:30Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.048002 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:30Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.085347 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:30Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.128215 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:30Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.137254 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.137357 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.137379 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:36:30 crc kubenswrapper[4787]: E1001 09:36:30.137433 4787 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 09:36:30 crc kubenswrapper[4787]: E1001 09:36:30.137435 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:36:32.137410485 +0000 UTC m=+24.252554642 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:36:30 crc kubenswrapper[4787]: E1001 09:36:30.137483 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 09:36:32.137474565 +0000 UTC m=+24.252618722 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.137577 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.137622 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:36:30 crc kubenswrapper[4787]: E1001 09:36:30.137690 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 09:36:30 crc kubenswrapper[4787]: E1001 09:36:30.137719 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 09:36:30 crc kubenswrapper[4787]: E1001 09:36:30.137732 4787 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:36:30 crc kubenswrapper[4787]: E1001 09:36:30.137785 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-01 09:36:32.137770177 +0000 UTC m=+24.252914334 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:36:30 crc kubenswrapper[4787]: E1001 09:36:30.137788 4787 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 09:36:30 crc kubenswrapper[4787]: E1001 09:36:30.137621 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 09:36:30 crc kubenswrapper[4787]: E1001 09:36:30.137819 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 09:36:30 crc kubenswrapper[4787]: E1001 09:36:30.137828 4787 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:36:30 crc kubenswrapper[4787]: E1001 09:36:30.137829 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 09:36:32.137822428 +0000 UTC m=+24.252966585 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 09:36:30 crc kubenswrapper[4787]: E1001 09:36:30.137863 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-01 09:36:32.137856268 +0000 UTC m=+24.253000425 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.161693 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:30Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.203427 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b093e96a244fdec4c5cdce62650d4da31c2ec46207a9c08bbd0a28bf286abb52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4664cb1dde25ada88989e81f0319aedc6dbd363f608a50623efaa3eeb8d00763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:30Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.240502 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d95f8b1aac879d1746d8a8af7e1d5e672b36c31d5b95c23fee70b237bc3716e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:30Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.283563 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0814bed7-1117-4dff-97ef-a3cd1240e615\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c91542e7254c8289b734b3ffa8e8226f2445c915c264e09c179b99604abad1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:36:27Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1001 09:36:22.478145 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 09:36:22.478910 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3375976971/tls.crt::/tmp/serving-cert-3375976971/tls.key\\\\\\\"\\\\nI1001 09:36:27.943129 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 09:36:27.955126 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 09:36:27.955153 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 09:36:27.955179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 09:36:27.955185 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 09:36:27.963955 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1001 09:36:27.963979 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1001 09:36:27.963986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964008 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 09:36:27.964015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 09:36:27.964018 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 09:36:27.964022 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1001 09:36:27.966112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:30Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.323256 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7aa5c938897ea621fc7c71c86e6a1dd2acca273e13afe71e9a27fc81f34cf2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:30Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.371282 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:30Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.408549 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:30Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.443985 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72701c6061105b7bb81545b5f0b035b220b22b2f8009e82809a037870c625685\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:30Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.485494 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:30Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.522449 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:30Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.523003 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.523120 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:36:30 crc kubenswrapper[4787]: E1001 09:36:30.523128 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:36:30 crc kubenswrapper[4787]: E1001 09:36:30.523333 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.527941 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.528660 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.529755 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.530473 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.531450 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.532010 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.561860 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:30Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.604558 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebc8a6e2b9c6104b0f773e99d68a62e1b2bf0adc05e6f291cf91e7a8cf1bc121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:30Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.664561 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" event={"ID":"eadf2545-886e-4642-b909-704c6cd4134e","Type":"ContainerStarted","Data":"e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4"} Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.664607 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" event={"ID":"eadf2545-886e-4642-b909-704c6cd4134e","Type":"ContainerStarted","Data":"834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e"} Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.664620 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" event={"ID":"eadf2545-886e-4642-b909-704c6cd4134e","Type":"ContainerStarted","Data":"39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274"} Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.664631 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" event={"ID":"eadf2545-886e-4642-b909-704c6cd4134e","Type":"ContainerStarted","Data":"7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f"} Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.664642 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" event={"ID":"eadf2545-886e-4642-b909-704c6cd4134e","Type":"ContainerStarted","Data":"b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a"} Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.664651 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" event={"ID":"eadf2545-886e-4642-b909-704c6cd4134e","Type":"ContainerStarted","Data":"394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694"} Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.666568 4787 generic.go:334] "Generic (PLEG): container finished" podID="60771b6a-9398-4987-bb28-a20ceb6cca38" containerID="58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26" exitCode=0 Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.667201 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" event={"ID":"60771b6a-9398-4987-bb28-a20ceb6cca38","Type":"ContainerDied","Data":"58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26"} Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.710065 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0814bed7-1117-4dff-97ef-a3cd1240e615\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c91542e7254c8289b734b3ffa8e8226f2445c915c264e09c179b99604abad1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:36:27Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1001 09:36:22.478145 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 09:36:22.478910 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3375976971/tls.crt::/tmp/serving-cert-3375976971/tls.key\\\\\\\"\\\\nI1001 09:36:27.943129 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 09:36:27.955126 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 09:36:27.955153 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 09:36:27.955179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 09:36:27.955185 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 09:36:27.963955 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1001 09:36:27.963979 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1001 09:36:27.963986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964008 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 09:36:27.964015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 09:36:27.964018 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 09:36:27.964022 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1001 09:36:27.966112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:30Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.733588 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7aa5c938897ea621fc7c71c86e6a1dd2acca273e13afe71e9a27fc81f34cf2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:30Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.752422 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:30Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.774741 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:30Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.802018 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:30Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.843278 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72701c6061105b7bb81545b5f0b035b220b22b2f8009e82809a037870c625685\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:30Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.882831 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:30Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.922090 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:30Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:30 crc kubenswrapper[4787]: I1001 09:36:30.964829 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebc8a6e2b9c6104b0f773e99d68a62e1b2bf0adc05e6f291cf91e7a8cf1bc121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:30Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:31 crc kubenswrapper[4787]: I1001 09:36:31.003922 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b093e96a244fdec4c5cdce62650d4da31c2ec46207a9c08bbd0a28bf286abb52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4664cb1dde25ada88989e81f0319aedc6dbd363f608a50623efaa3eeb8d00763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:31Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:31 crc kubenswrapper[4787]: I1001 09:36:31.043241 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d95f8b1aac879d1746d8a8af7e1d5e672b36c31d5b95c23fee70b237bc3716e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:31Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:31 crc kubenswrapper[4787]: I1001 09:36:31.088934 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:31Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:31 crc kubenswrapper[4787]: I1001 09:36:31.121429 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:31Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:31 crc kubenswrapper[4787]: I1001 09:36:31.523539 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:36:31 crc kubenswrapper[4787]: E1001 09:36:31.523670 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:36:31 crc kubenswrapper[4787]: I1001 09:36:31.673708 4787 generic.go:334] "Generic (PLEG): container finished" podID="60771b6a-9398-4987-bb28-a20ceb6cca38" containerID="3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81" exitCode=0 Oct 01 09:36:31 crc kubenswrapper[4787]: I1001 09:36:31.673806 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" event={"ID":"60771b6a-9398-4987-bb28-a20ceb6cca38","Type":"ContainerDied","Data":"3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81"} Oct 01 09:36:31 crc kubenswrapper[4787]: I1001 09:36:31.692144 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7aa5c938897ea621fc7c71c86e6a1dd2acca273e13afe71e9a27fc81f34cf2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:31Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:31 crc kubenswrapper[4787]: I1001 09:36:31.710817 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:31Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:31 crc kubenswrapper[4787]: I1001 09:36:31.736409 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:31Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:31 crc kubenswrapper[4787]: I1001 09:36:31.755386 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0814bed7-1117-4dff-97ef-a3cd1240e615\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c91542e7254c8289b734b3ffa8e8226f2445c915c264e09c179b99604abad1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:36:27Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1001 09:36:22.478145 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 09:36:22.478910 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3375976971/tls.crt::/tmp/serving-cert-3375976971/tls.key\\\\\\\"\\\\nI1001 09:36:27.943129 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 09:36:27.955126 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 09:36:27.955153 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 09:36:27.955179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 09:36:27.955185 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 09:36:27.963955 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1001 09:36:27.963979 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1001 09:36:27.963986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964008 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 09:36:27.964015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 09:36:27.964018 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 09:36:27.964022 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1001 09:36:27.966112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:31Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:31 crc kubenswrapper[4787]: I1001 09:36:31.772273 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72701c6061105b7bb81545b5f0b035b220b22b2f8009e82809a037870c625685\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:31Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:31 crc kubenswrapper[4787]: I1001 09:36:31.786263 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:31Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:31 crc kubenswrapper[4787]: I1001 09:36:31.800600 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:31Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:31 crc kubenswrapper[4787]: I1001 09:36:31.813564 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebc8a6e2b9c6104b0f773e99d68a62e1b2bf0adc05e6f291cf91e7a8cf1bc121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:31Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:31 crc kubenswrapper[4787]: I1001 09:36:31.834567 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:31Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:31 crc kubenswrapper[4787]: I1001 09:36:31.851354 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:31Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:31 crc kubenswrapper[4787]: I1001 09:36:31.867696 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:31Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:31 crc kubenswrapper[4787]: I1001 09:36:31.881614 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b093e96a244fdec4c5cdce62650d4da31c2ec46207a9c08bbd0a28bf286abb52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4664cb1dde25ada88989e81f0319aedc6dbd363f608a50623efaa3eeb8d00763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:31Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:31 crc kubenswrapper[4787]: I1001 09:36:31.894342 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d95f8b1aac879d1746d8a8af7e1d5e672b36c31d5b95c23fee70b237bc3716e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:31Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:32 crc kubenswrapper[4787]: I1001 09:36:32.158349 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:36:32 crc kubenswrapper[4787]: I1001 09:36:32.158444 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:36:32 crc kubenswrapper[4787]: I1001 09:36:32.158489 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:36:32 crc kubenswrapper[4787]: E1001 09:36:32.158514 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:36:36.158498238 +0000 UTC m=+28.273642385 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:36:32 crc kubenswrapper[4787]: I1001 09:36:32.158540 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:36:32 crc kubenswrapper[4787]: I1001 09:36:32.158568 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:36:32 crc kubenswrapper[4787]: E1001 09:36:32.158619 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 09:36:32 crc kubenswrapper[4787]: E1001 09:36:32.158642 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 09:36:32 crc kubenswrapper[4787]: E1001 09:36:32.158677 4787 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:36:32 crc kubenswrapper[4787]: E1001 09:36:32.158708 4787 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 09:36:32 crc kubenswrapper[4787]: E1001 09:36:32.158733 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-01 09:36:36.158719584 +0000 UTC m=+28.273863751 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:36:32 crc kubenswrapper[4787]: E1001 09:36:32.158643 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 09:36:32 crc kubenswrapper[4787]: E1001 09:36:32.158765 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 09:36:32 crc kubenswrapper[4787]: E1001 09:36:32.158775 4787 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:36:32 crc kubenswrapper[4787]: E1001 09:36:32.158799 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 09:36:36.158771985 +0000 UTC m=+28.273916142 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 09:36:32 crc kubenswrapper[4787]: E1001 09:36:32.158820 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-01 09:36:36.158812676 +0000 UTC m=+28.273956833 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:36:32 crc kubenswrapper[4787]: E1001 09:36:32.158832 4787 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 09:36:32 crc kubenswrapper[4787]: E1001 09:36:32.159039 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 09:36:36.1589756 +0000 UTC m=+28.274119797 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 09:36:32 crc kubenswrapper[4787]: I1001 09:36:32.522860 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:36:32 crc kubenswrapper[4787]: E1001 09:36:32.523010 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:36:32 crc kubenswrapper[4787]: I1001 09:36:32.523069 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:36:32 crc kubenswrapper[4787]: E1001 09:36:32.523283 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:36:32 crc kubenswrapper[4787]: I1001 09:36:32.681842 4787 generic.go:334] "Generic (PLEG): container finished" podID="60771b6a-9398-4987-bb28-a20ceb6cca38" containerID="ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5" exitCode=0 Oct 01 09:36:32 crc kubenswrapper[4787]: I1001 09:36:32.681973 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" event={"ID":"60771b6a-9398-4987-bb28-a20ceb6cca38","Type":"ContainerDied","Data":"ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5"} Oct 01 09:36:32 crc kubenswrapper[4787]: I1001 09:36:32.688683 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" event={"ID":"eadf2545-886e-4642-b909-704c6cd4134e","Type":"ContainerStarted","Data":"85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68"} Oct 01 09:36:32 crc kubenswrapper[4787]: I1001 09:36:32.690814 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"f4990da342a20f6ab1e97774294bd6856c49ca17de21d893a638c13c8e727dba"} Oct 01 09:36:32 crc kubenswrapper[4787]: I1001 09:36:32.708164 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72701c6061105b7bb81545b5f0b035b220b22b2f8009e82809a037870c625685\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:32Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:32 crc kubenswrapper[4787]: I1001 09:36:32.725611 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:32Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:32 crc kubenswrapper[4787]: I1001 09:36:32.742263 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:32Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:32 crc kubenswrapper[4787]: I1001 09:36:32.758539 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:32Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:32 crc kubenswrapper[4787]: I1001 09:36:32.771115 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebc8a6e2b9c6104b0f773e99d68a62e1b2bf0adc05e6f291cf91e7a8cf1bc121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:32Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:32 crc kubenswrapper[4787]: I1001 09:36:32.784883 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:32Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:32 crc kubenswrapper[4787]: I1001 09:36:32.800278 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:32Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:32 crc kubenswrapper[4787]: I1001 09:36:32.813500 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b093e96a244fdec4c5cdce62650d4da31c2ec46207a9c08bbd0a28bf286abb52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4664cb1dde25ada88989e81f0319aedc6dbd363f608a50623efaa3eeb8d00763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:32Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:32 crc kubenswrapper[4787]: I1001 09:36:32.825213 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d95f8b1aac879d1746d8a8af7e1d5e672b36c31d5b95c23fee70b237bc3716e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:32Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:32 crc kubenswrapper[4787]: I1001 09:36:32.844526 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:32Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:32 crc kubenswrapper[4787]: I1001 09:36:32.877512 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:32Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:32 crc kubenswrapper[4787]: I1001 09:36:32.894880 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0814bed7-1117-4dff-97ef-a3cd1240e615\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c91542e7254c8289b734b3ffa8e8226f2445c915c264e09c179b99604abad1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:36:27Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1001 09:36:22.478145 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 09:36:22.478910 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3375976971/tls.crt::/tmp/serving-cert-3375976971/tls.key\\\\\\\"\\\\nI1001 09:36:27.943129 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 09:36:27.955126 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 09:36:27.955153 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 09:36:27.955179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 09:36:27.955185 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 09:36:27.963955 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1001 09:36:27.963979 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1001 09:36:27.963986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964008 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 09:36:27.964015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 09:36:27.964018 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 09:36:27.964022 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1001 09:36:27.966112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:32Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:32 crc kubenswrapper[4787]: I1001 09:36:32.910215 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7aa5c938897ea621fc7c71c86e6a1dd2acca273e13afe71e9a27fc81f34cf2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:32Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:32 crc kubenswrapper[4787]: I1001 09:36:32.925658 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:32Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:32 crc kubenswrapper[4787]: I1001 09:36:32.937796 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebc8a6e2b9c6104b0f773e99d68a62e1b2bf0adc05e6f291cf91e7a8cf1bc121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:32Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:32 crc kubenswrapper[4787]: I1001 09:36:32.949187 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:32Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:32 crc kubenswrapper[4787]: I1001 09:36:32.963009 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4990da342a20f6ab1e97774294bd6856c49ca17de21d893a638c13c8e727dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:32Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:32 crc kubenswrapper[4787]: I1001 09:36:32.975966 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b093e96a244fdec4c5cdce62650d4da31c2ec46207a9c08bbd0a28bf286abb52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4664cb1dde25ada88989e81f0319aedc6dbd363f608a50623efaa3eeb8d00763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:32Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:32 crc kubenswrapper[4787]: I1001 09:36:32.988392 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d95f8b1aac879d1746d8a8af7e1d5e672b36c31d5b95c23fee70b237bc3716e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:32Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.000770 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:32Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.016687 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.032347 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0814bed7-1117-4dff-97ef-a3cd1240e615\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c91542e7254c8289b734b3ffa8e8226f2445c915c264e09c179b99604abad1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:36:27Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1001 09:36:22.478145 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 09:36:22.478910 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3375976971/tls.crt::/tmp/serving-cert-3375976971/tls.key\\\\\\\"\\\\nI1001 09:36:27.943129 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 09:36:27.955126 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 09:36:27.955153 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 09:36:27.955179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 09:36:27.955185 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 09:36:27.963955 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1001 09:36:27.963979 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1001 09:36:27.963986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964008 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 09:36:27.964015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 09:36:27.964018 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 09:36:27.964022 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1001 09:36:27.966112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.045540 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7aa5c938897ea621fc7c71c86e6a1dd2acca273e13afe71e9a27fc81f34cf2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.057677 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72701c6061105b7bb81545b5f0b035b220b22b2f8009e82809a037870c625685\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.071249 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.085386 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.210670 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.216014 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.223668 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.227043 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.243686 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.257272 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72701c6061105b7bb81545b5f0b035b220b22b2f8009e82809a037870c625685\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.269948 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.280850 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebc8a6e2b9c6104b0f773e99d68a62e1b2bf0adc05e6f291cf91e7a8cf1bc121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.291951 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4990da342a20f6ab1e97774294bd6856c49ca17de21d893a638c13c8e727dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.308136 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b093e96a244fdec4c5cdce62650d4da31c2ec46207a9c08bbd0a28bf286abb52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4664cb1dde25ada88989e81f0319aedc6dbd363f608a50623efaa3eeb8d00763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.320294 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d95f8b1aac879d1746d8a8af7e1d5e672b36c31d5b95c23fee70b237bc3716e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.330967 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.350358 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.370828 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0814bed7-1117-4dff-97ef-a3cd1240e615\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c91542e7254c8289b734b3ffa8e8226f2445c915c264e09c179b99604abad1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:36:27Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1001 09:36:22.478145 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 09:36:22.478910 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3375976971/tls.crt::/tmp/serving-cert-3375976971/tls.key\\\\\\\"\\\\nI1001 09:36:27.943129 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 09:36:27.955126 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 09:36:27.955153 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 09:36:27.955179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 09:36:27.955185 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 09:36:27.963955 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1001 09:36:27.963979 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1001 09:36:27.963986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964008 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 09:36:27.964015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 09:36:27.964018 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 09:36:27.964022 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1001 09:36:27.966112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.383925 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7aa5c938897ea621fc7c71c86e6a1dd2acca273e13afe71e9a27fc81f34cf2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.397517 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.410236 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.420363 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebc8a6e2b9c6104b0f773e99d68a62e1b2bf0adc05e6f291cf91e7a8cf1bc121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.433206 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.454773 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4990da342a20f6ab1e97774294bd6856c49ca17de21d893a638c13c8e727dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.478029 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b093e96a244fdec4c5cdce62650d4da31c2ec46207a9c08bbd0a28bf286abb52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4664cb1dde25ada88989e81f0319aedc6dbd363f608a50623efaa3eeb8d00763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.494631 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d95f8b1aac879d1746d8a8af7e1d5e672b36c31d5b95c23fee70b237bc3716e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.515261 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0814bed7-1117-4dff-97ef-a3cd1240e615\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c91542e7254c8289b734b3ffa8e8226f2445c915c264e09c179b99604abad1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:36:27Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1001 09:36:22.478145 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 09:36:22.478910 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3375976971/tls.crt::/tmp/serving-cert-3375976971/tls.key\\\\\\\"\\\\nI1001 09:36:27.943129 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 09:36:27.955126 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 09:36:27.955153 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 09:36:27.955179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 09:36:27.955185 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 09:36:27.963955 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1001 09:36:27.963979 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1001 09:36:27.963986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964008 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 09:36:27.964015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 09:36:27.964018 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 09:36:27.964022 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1001 09:36:27.966112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.522841 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:36:33 crc kubenswrapper[4787]: E1001 09:36:33.522964 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.538882 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7aa5c938897ea621fc7c71c86e6a1dd2acca273e13afe71e9a27fc81f34cf2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.582437 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.629626 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.662007 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc37698-af76-4818-b216-ffe35af67929\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1bb16edb25b63e957cc0bceb04a9fa9244e68881cb45be7c7d627c5faa74232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://995030421e0cbfeb7baa8ccb574d269dd178f8185bcaa9000fe5e7fc221b1188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d4c78dd607b81a28bcfe0ffa69c02628cf06aff5fd6a574069639c5bac044a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://144123d273f3f1af4fb857d731a16d9891b097c98229a035eb132f360fa8b2f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.696174 4787 generic.go:334] "Generic (PLEG): container finished" podID="60771b6a-9398-4987-bb28-a20ceb6cca38" containerID="d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54" exitCode=0 Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.696264 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" event={"ID":"60771b6a-9398-4987-bb28-a20ceb6cca38","Type":"ContainerDied","Data":"d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54"} Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.702369 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72701c6061105b7bb81545b5f0b035b220b22b2f8009e82809a037870c625685\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.740495 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.782590 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.823097 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.861222 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebc8a6e2b9c6104b0f773e99d68a62e1b2bf0adc05e6f291cf91e7a8cf1bc121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.905137 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.942638 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4990da342a20f6ab1e97774294bd6856c49ca17de21d893a638c13c8e727dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:33 crc kubenswrapper[4787]: I1001 09:36:33.982938 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b093e96a244fdec4c5cdce62650d4da31c2ec46207a9c08bbd0a28bf286abb52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4664cb1dde25ada88989e81f0319aedc6dbd363f608a50623efaa3eeb8d00763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:33Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.022734 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d95f8b1aac879d1746d8a8af7e1d5e672b36c31d5b95c23fee70b237bc3716e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:34Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.064001 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0814bed7-1117-4dff-97ef-a3cd1240e615\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c91542e7254c8289b734b3ffa8e8226f2445c915c264e09c179b99604abad1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:36:27Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1001 09:36:22.478145 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 09:36:22.478910 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3375976971/tls.crt::/tmp/serving-cert-3375976971/tls.key\\\\\\\"\\\\nI1001 09:36:27.943129 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 09:36:27.955126 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 09:36:27.955153 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 09:36:27.955179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 09:36:27.955185 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 09:36:27.963955 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1001 09:36:27.963979 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1001 09:36:27.963986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964008 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 09:36:27.964015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 09:36:27.964018 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 09:36:27.964022 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1001 09:36:27.966112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:34Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.102230 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7aa5c938897ea621fc7c71c86e6a1dd2acca273e13afe71e9a27fc81f34cf2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:34Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.146788 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:34Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.188114 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:34Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.221369 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc37698-af76-4818-b216-ffe35af67929\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1bb16edb25b63e957cc0bceb04a9fa9244e68881cb45be7c7d627c5faa74232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://995030421e0cbfeb7baa8ccb574d269dd178f8185bcaa9000fe5e7fc221b1188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d4c78dd607b81a28bcfe0ffa69c02628cf06aff5fd6a574069639c5bac044a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://144123d273f3f1af4fb857d731a16d9891b097c98229a035eb132f360fa8b2f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:34Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.262609 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72701c6061105b7bb81545b5f0b035b220b22b2f8009e82809a037870c625685\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:34Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.283420 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.285298 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.285346 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.285364 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.285458 4787 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.307882 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:34Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.355282 4787 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.355681 4787 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.357323 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.357380 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.357401 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.357430 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.357448 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:34Z","lastTransitionTime":"2025-10-01T09:36:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:34 crc kubenswrapper[4787]: E1001 09:36:34.375401 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:34Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.379384 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.379425 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.379439 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.379458 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.379472 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:34Z","lastTransitionTime":"2025-10-01T09:36:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.385317 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:34Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:34 crc kubenswrapper[4787]: E1001 09:36:34.395511 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:34Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.401345 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.401527 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.401625 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.401715 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.402006 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:34Z","lastTransitionTime":"2025-10-01T09:36:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:34 crc kubenswrapper[4787]: E1001 09:36:34.418435 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:34Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.423327 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.423364 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.423378 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.423400 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.423415 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:34Z","lastTransitionTime":"2025-10-01T09:36:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:34 crc kubenswrapper[4787]: E1001 09:36:34.438261 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:34Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.441422 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.441569 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.441735 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.441968 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.442198 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:34Z","lastTransitionTime":"2025-10-01T09:36:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:34 crc kubenswrapper[4787]: E1001 09:36:34.456806 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:34Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:34 crc kubenswrapper[4787]: E1001 09:36:34.457240 4787 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.459369 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.459479 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.459565 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.459682 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.459768 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:34Z","lastTransitionTime":"2025-10-01T09:36:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.523177 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.523246 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:36:34 crc kubenswrapper[4787]: E1001 09:36:34.523543 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:36:34 crc kubenswrapper[4787]: E1001 09:36:34.523692 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.562683 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.563110 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.563195 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.563278 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.563353 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:34Z","lastTransitionTime":"2025-10-01T09:36:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.666357 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.666611 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.666709 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.666787 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.666842 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:34Z","lastTransitionTime":"2025-10-01T09:36:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.706267 4787 generic.go:334] "Generic (PLEG): container finished" podID="60771b6a-9398-4987-bb28-a20ceb6cca38" containerID="3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b" exitCode=0 Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.706367 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" event={"ID":"60771b6a-9398-4987-bb28-a20ceb6cca38","Type":"ContainerDied","Data":"3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b"} Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.725161 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d95f8b1aac879d1746d8a8af7e1d5e672b36c31d5b95c23fee70b237bc3716e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:34Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.744032 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:34Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.759016 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4990da342a20f6ab1e97774294bd6856c49ca17de21d893a638c13c8e727dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:34Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.769190 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.769228 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.769239 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.769252 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.769263 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:34Z","lastTransitionTime":"2025-10-01T09:36:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.776473 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b093e96a244fdec4c5cdce62650d4da31c2ec46207a9c08bbd0a28bf286abb52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4664cb1dde25ada88989e81f0319aedc6dbd363f608a50623efaa3eeb8d00763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:34Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.799428 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0814bed7-1117-4dff-97ef-a3cd1240e615\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c91542e7254c8289b734b3ffa8e8226f2445c915c264e09c179b99604abad1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:36:27Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1001 09:36:22.478145 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 09:36:22.478910 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3375976971/tls.crt::/tmp/serving-cert-3375976971/tls.key\\\\\\\"\\\\nI1001 09:36:27.943129 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 09:36:27.955126 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 09:36:27.955153 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 09:36:27.955179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 09:36:27.955185 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 09:36:27.963955 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1001 09:36:27.963979 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1001 09:36:27.963986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964008 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 09:36:27.964015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 09:36:27.964018 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 09:36:27.964022 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1001 09:36:27.966112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:34Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.813164 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7aa5c938897ea621fc7c71c86e6a1dd2acca273e13afe71e9a27fc81f34cf2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:34Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.834786 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:34Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.854297 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:34Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.872323 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.872362 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.872373 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.872389 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.872402 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:34Z","lastTransitionTime":"2025-10-01T09:36:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.877484 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc37698-af76-4818-b216-ffe35af67929\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1bb16edb25b63e957cc0bceb04a9fa9244e68881cb45be7c7d627c5faa74232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://995030421e0cbfeb7baa8ccb574d269dd178f8185bcaa9000fe5e7fc221b1188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d4c78dd607b81a28bcfe0ffa69c02628cf06aff5fd6a574069639c5bac044a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://144123d273f3f1af4fb857d731a16d9891b097c98229a035eb132f360fa8b2f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:34Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.892671 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72701c6061105b7bb81545b5f0b035b220b22b2f8009e82809a037870c625685\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:34Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.906812 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:34Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.921332 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:34Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.940258 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:34Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.954607 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebc8a6e2b9c6104b0f773e99d68a62e1b2bf0adc05e6f291cf91e7a8cf1bc121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:34Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.975029 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.975086 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.975100 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.975123 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:34 crc kubenswrapper[4787]: I1001 09:36:34.975132 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:34Z","lastTransitionTime":"2025-10-01T09:36:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.078826 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.078883 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.078900 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.078926 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.078946 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:35Z","lastTransitionTime":"2025-10-01T09:36:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.182814 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.182860 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.182873 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.182892 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.182905 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:35Z","lastTransitionTime":"2025-10-01T09:36:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.286996 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.287031 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.287043 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.287060 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.287071 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:35Z","lastTransitionTime":"2025-10-01T09:36:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.390650 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.390725 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.390744 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.390774 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.390794 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:35Z","lastTransitionTime":"2025-10-01T09:36:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.493516 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.493565 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.493576 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.493592 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.493603 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:35Z","lastTransitionTime":"2025-10-01T09:36:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.523466 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:36:35 crc kubenswrapper[4787]: E1001 09:36:35.523652 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.595787 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.595841 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.595851 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.595871 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.595884 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:35Z","lastTransitionTime":"2025-10-01T09:36:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.700113 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.700199 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.700213 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.700238 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.700253 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:35Z","lastTransitionTime":"2025-10-01T09:36:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.719775 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" event={"ID":"eadf2545-886e-4642-b909-704c6cd4134e","Type":"ContainerStarted","Data":"72de5132325b6af956d84387ae27126362df626f66d135a50e420fc28bc0edc0"} Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.720496 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.720555 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.726690 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" event={"ID":"60771b6a-9398-4987-bb28-a20ceb6cca38","Type":"ContainerStarted","Data":"c68b374f72c2a689eb09480e7ae3e88691daf0cb380a52f655eb1c651a66a281"} Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.740384 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:35Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.803485 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.803715 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.803724 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.803738 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.803746 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:35Z","lastTransitionTime":"2025-10-01T09:36:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.810776 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.811619 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72de5132325b6af956d84387ae27126362df626f66d135a50e420fc28bc0edc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:35Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.812159 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.830162 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0814bed7-1117-4dff-97ef-a3cd1240e615\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c91542e7254c8289b734b3ffa8e8226f2445c915c264e09c179b99604abad1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:36:27Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1001 09:36:22.478145 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 09:36:22.478910 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3375976971/tls.crt::/tmp/serving-cert-3375976971/tls.key\\\\\\\"\\\\nI1001 09:36:27.943129 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 09:36:27.955126 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 09:36:27.955153 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 09:36:27.955179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 09:36:27.955185 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 09:36:27.963955 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1001 09:36:27.963979 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1001 09:36:27.963986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964008 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 09:36:27.964015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 09:36:27.964018 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 09:36:27.964022 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1001 09:36:27.966112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:35Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.844309 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7aa5c938897ea621fc7c71c86e6a1dd2acca273e13afe71e9a27fc81f34cf2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:35Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.869164 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72701c6061105b7bb81545b5f0b035b220b22b2f8009e82809a037870c625685\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:35Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.889058 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:35Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.906218 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.906297 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.906312 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.906338 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.906358 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:35Z","lastTransitionTime":"2025-10-01T09:36:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.912872 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:35Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.929607 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc37698-af76-4818-b216-ffe35af67929\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1bb16edb25b63e957cc0bceb04a9fa9244e68881cb45be7c7d627c5faa74232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://995030421e0cbfeb7baa8ccb574d269dd178f8185bcaa9000fe5e7fc221b1188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d4c78dd607b81a28bcfe0ffa69c02628cf06aff5fd6a574069639c5bac044a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://144123d273f3f1af4fb857d731a16d9891b097c98229a035eb132f360fa8b2f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:35Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.943287 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:35Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.957809 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebc8a6e2b9c6104b0f773e99d68a62e1b2bf0adc05e6f291cf91e7a8cf1bc121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:35Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.975001 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:35Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:35 crc kubenswrapper[4787]: I1001 09:36:35.987169 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4990da342a20f6ab1e97774294bd6856c49ca17de21d893a638c13c8e727dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:35Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.002177 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b093e96a244fdec4c5cdce62650d4da31c2ec46207a9c08bbd0a28bf286abb52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4664cb1dde25ada88989e81f0319aedc6dbd363f608a50623efaa3eeb8d00763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:36Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.009831 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.009863 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.009873 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.009889 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.009899 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:36Z","lastTransitionTime":"2025-10-01T09:36:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.018167 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d95f8b1aac879d1746d8a8af7e1d5e672b36c31d5b95c23fee70b237bc3716e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:36Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.035263 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0814bed7-1117-4dff-97ef-a3cd1240e615\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c91542e7254c8289b734b3ffa8e8226f2445c915c264e09c179b99604abad1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:36:27Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1001 09:36:22.478145 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 09:36:22.478910 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3375976971/tls.crt::/tmp/serving-cert-3375976971/tls.key\\\\\\\"\\\\nI1001 09:36:27.943129 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 09:36:27.955126 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 09:36:27.955153 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 09:36:27.955179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 09:36:27.955185 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 09:36:27.963955 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1001 09:36:27.963979 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1001 09:36:27.963986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964008 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 09:36:27.964015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 09:36:27.964018 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 09:36:27.964022 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1001 09:36:27.966112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:36Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.049041 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7aa5c938897ea621fc7c71c86e6a1dd2acca273e13afe71e9a27fc81f34cf2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:36Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.074371 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c68b374f72c2a689eb09480e7ae3e88691daf0cb380a52f655eb1c651a66a281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:36Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.107777 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72de5132325b6af956d84387ae27126362df626f66d135a50e420fc28bc0edc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:36Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.113757 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.113811 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.113824 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.113847 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.113861 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:36Z","lastTransitionTime":"2025-10-01T09:36:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.125950 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:36Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.142434 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc37698-af76-4818-b216-ffe35af67929\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1bb16edb25b63e957cc0bceb04a9fa9244e68881cb45be7c7d627c5faa74232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://995030421e0cbfeb7baa8ccb574d269dd178f8185bcaa9000fe5e7fc221b1188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d4c78dd607b81a28bcfe0ffa69c02628cf06aff5fd6a574069639c5bac044a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://144123d273f3f1af4fb857d731a16d9891b097c98229a035eb132f360fa8b2f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:36Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.159031 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72701c6061105b7bb81545b5f0b035b220b22b2f8009e82809a037870c625685\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:36Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.179343 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:36Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.194105 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:36Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.210285 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebc8a6e2b9c6104b0f773e99d68a62e1b2bf0adc05e6f291cf91e7a8cf1bc121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:36Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.212676 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.212819 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:36:36 crc kubenswrapper[4787]: E1001 09:36:36.212848 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:36:44.212815745 +0000 UTC m=+36.327959912 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.212899 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:36:36 crc kubenswrapper[4787]: E1001 09:36:36.212978 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 09:36:36 crc kubenswrapper[4787]: E1001 09:36:36.213024 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 09:36:36 crc kubenswrapper[4787]: E1001 09:36:36.213047 4787 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:36:36 crc kubenswrapper[4787]: E1001 09:36:36.213061 4787 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 09:36:36 crc kubenswrapper[4787]: E1001 09:36:36.213144 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-01 09:36:44.213118862 +0000 UTC m=+36.328263209 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:36:36 crc kubenswrapper[4787]: E1001 09:36:36.213190 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 09:36:44.213165453 +0000 UTC m=+36.328309610 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 09:36:36 crc kubenswrapper[4787]: E1001 09:36:36.213190 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 09:36:36 crc kubenswrapper[4787]: E1001 09:36:36.213228 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 09:36:36 crc kubenswrapper[4787]: E1001 09:36:36.213244 4787 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:36:36 crc kubenswrapper[4787]: E1001 09:36:36.213295 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-01 09:36:44.213280076 +0000 UTC m=+36.328424253 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.212987 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.213372 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:36:36 crc kubenswrapper[4787]: E1001 09:36:36.213497 4787 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 09:36:36 crc kubenswrapper[4787]: E1001 09:36:36.213542 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 09:36:44.213532363 +0000 UTC m=+36.328676520 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.216782 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.216876 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.216951 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.217015 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.217110 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:36Z","lastTransitionTime":"2025-10-01T09:36:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.227214 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b093e96a244fdec4c5cdce62650d4da31c2ec46207a9c08bbd0a28bf286abb52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4664cb1dde25ada88989e81f0319aedc6dbd363f608a50623efaa3eeb8d00763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:36Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.245209 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d95f8b1aac879d1746d8a8af7e1d5e672b36c31d5b95c23fee70b237bc3716e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:36Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.264244 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:36Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.284513 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4990da342a20f6ab1e97774294bd6856c49ca17de21d893a638c13c8e727dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:36Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.320532 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.321251 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.321320 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.321371 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.321394 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:36Z","lastTransitionTime":"2025-10-01T09:36:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.425286 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.425364 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.425385 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.425417 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.425439 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:36Z","lastTransitionTime":"2025-10-01T09:36:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.523293 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:36:36 crc kubenswrapper[4787]: E1001 09:36:36.523439 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.523298 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:36:36 crc kubenswrapper[4787]: E1001 09:36:36.523592 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.528149 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.528226 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.528242 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.528259 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.528273 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:36Z","lastTransitionTime":"2025-10-01T09:36:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.631416 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.631485 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.631499 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.631519 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.631533 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:36Z","lastTransitionTime":"2025-10-01T09:36:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.729891 4787 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.734515 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.734600 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.734622 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.734656 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.734676 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:36Z","lastTransitionTime":"2025-10-01T09:36:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.838364 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.838787 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.838899 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.839026 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.839177 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:36Z","lastTransitionTime":"2025-10-01T09:36:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.941838 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.941872 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.941881 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.941896 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:36 crc kubenswrapper[4787]: I1001 09:36:36.941908 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:36Z","lastTransitionTime":"2025-10-01T09:36:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.044997 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.045066 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.045131 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.045160 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.045178 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:37Z","lastTransitionTime":"2025-10-01T09:36:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.147939 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.148000 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.148014 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.148043 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.148059 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:37Z","lastTransitionTime":"2025-10-01T09:36:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.250535 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.250572 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.250580 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.250593 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.250602 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:37Z","lastTransitionTime":"2025-10-01T09:36:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.352897 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.352979 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.352997 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.353018 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.353032 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:37Z","lastTransitionTime":"2025-10-01T09:36:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.455479 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.455617 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.455678 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.455744 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.455800 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:37Z","lastTransitionTime":"2025-10-01T09:36:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.522931 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:36:37 crc kubenswrapper[4787]: E1001 09:36:37.523055 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.557538 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.557569 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.557577 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.557588 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.557607 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:37Z","lastTransitionTime":"2025-10-01T09:36:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.660924 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.661288 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.661405 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.661696 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.661795 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:37Z","lastTransitionTime":"2025-10-01T09:36:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.734176 4787 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.765652 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.765736 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.765754 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.765786 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.765813 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:37Z","lastTransitionTime":"2025-10-01T09:36:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.868287 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.868353 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.868370 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.868392 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.868406 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:37Z","lastTransitionTime":"2025-10-01T09:36:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.971128 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.971168 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.971181 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.971199 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:37 crc kubenswrapper[4787]: I1001 09:36:37.971210 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:37Z","lastTransitionTime":"2025-10-01T09:36:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.073532 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.073578 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.073593 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.073613 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.073628 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:38Z","lastTransitionTime":"2025-10-01T09:36:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.175933 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.176000 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.176021 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.176048 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.176070 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:38Z","lastTransitionTime":"2025-10-01T09:36:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.278523 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.278574 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.278585 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.278605 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.278617 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:38Z","lastTransitionTime":"2025-10-01T09:36:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.380945 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.380991 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.381001 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.381016 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.381028 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:38Z","lastTransitionTime":"2025-10-01T09:36:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.483944 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.483989 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.484000 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.484016 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.484030 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:38Z","lastTransitionTime":"2025-10-01T09:36:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.522818 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.522852 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:36:38 crc kubenswrapper[4787]: E1001 09:36:38.522968 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:36:38 crc kubenswrapper[4787]: E1001 09:36:38.523062 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.543760 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc37698-af76-4818-b216-ffe35af67929\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1bb16edb25b63e957cc0bceb04a9fa9244e68881cb45be7c7d627c5faa74232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://995030421e0cbfeb7baa8ccb574d269dd178f8185bcaa9000fe5e7fc221b1188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d4c78dd607b81a28bcfe0ffa69c02628cf06aff5fd6a574069639c5bac044a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://144123d273f3f1af4fb857d731a16d9891b097c98229a035eb132f360fa8b2f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.558939 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72701c6061105b7bb81545b5f0b035b220b22b2f8009e82809a037870c625685\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.580418 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.586938 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.587264 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.587420 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.587559 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.587725 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:38Z","lastTransitionTime":"2025-10-01T09:36:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.603663 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.623720 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.637838 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebc8a6e2b9c6104b0f773e99d68a62e1b2bf0adc05e6f291cf91e7a8cf1bc121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.660172 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d95f8b1aac879d1746d8a8af7e1d5e672b36c31d5b95c23fee70b237bc3716e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.682014 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.690552 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.690603 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.690616 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.690640 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.690654 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:38Z","lastTransitionTime":"2025-10-01T09:36:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.694742 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4990da342a20f6ab1e97774294bd6856c49ca17de21d893a638c13c8e727dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.708496 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b093e96a244fdec4c5cdce62650d4da31c2ec46207a9c08bbd0a28bf286abb52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4664cb1dde25ada88989e81f0319aedc6dbd363f608a50623efaa3eeb8d00763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.726023 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0814bed7-1117-4dff-97ef-a3cd1240e615\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c91542e7254c8289b734b3ffa8e8226f2445c915c264e09c179b99604abad1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:36:27Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1001 09:36:22.478145 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 09:36:22.478910 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3375976971/tls.crt::/tmp/serving-cert-3375976971/tls.key\\\\\\\"\\\\nI1001 09:36:27.943129 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 09:36:27.955126 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 09:36:27.955153 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 09:36:27.955179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 09:36:27.955185 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 09:36:27.963955 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1001 09:36:27.963979 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1001 09:36:27.963986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964008 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 09:36:27.964015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 09:36:27.964018 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 09:36:27.964022 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1001 09:36:27.966112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.739049 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7aa5c938897ea621fc7c71c86e6a1dd2acca273e13afe71e9a27fc81f34cf2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.739706 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dmsxm_eadf2545-886e-4642-b909-704c6cd4134e/ovnkube-controller/0.log" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.743059 4787 generic.go:334] "Generic (PLEG): container finished" podID="eadf2545-886e-4642-b909-704c6cd4134e" containerID="72de5132325b6af956d84387ae27126362df626f66d135a50e420fc28bc0edc0" exitCode=1 Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.743129 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" event={"ID":"eadf2545-886e-4642-b909-704c6cd4134e","Type":"ContainerDied","Data":"72de5132325b6af956d84387ae27126362df626f66d135a50e420fc28bc0edc0"} Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.744212 4787 scope.go:117] "RemoveContainer" containerID="72de5132325b6af956d84387ae27126362df626f66d135a50e420fc28bc0edc0" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.757187 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c68b374f72c2a689eb09480e7ae3e88691daf0cb380a52f655eb1c651a66a281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.775259 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72de5132325b6af956d84387ae27126362df626f66d135a50e420fc28bc0edc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.789962 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.803655 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.803693 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.803718 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.803736 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.803748 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:38Z","lastTransitionTime":"2025-10-01T09:36:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.826382 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4990da342a20f6ab1e97774294bd6856c49ca17de21d893a638c13c8e727dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.848953 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b093e96a244fdec4c5cdce62650d4da31c2ec46207a9c08bbd0a28bf286abb52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4664cb1dde25ada88989e81f0319aedc6dbd363f608a50623efaa3eeb8d00763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.861782 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d95f8b1aac879d1746d8a8af7e1d5e672b36c31d5b95c23fee70b237bc3716e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.878674 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0814bed7-1117-4dff-97ef-a3cd1240e615\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c91542e7254c8289b734b3ffa8e8226f2445c915c264e09c179b99604abad1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:36:27Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1001 09:36:22.478145 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 09:36:22.478910 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3375976971/tls.crt::/tmp/serving-cert-3375976971/tls.key\\\\\\\"\\\\nI1001 09:36:27.943129 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 09:36:27.955126 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 09:36:27.955153 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 09:36:27.955179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 09:36:27.955185 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 09:36:27.963955 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1001 09:36:27.963979 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1001 09:36:27.963986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964008 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 09:36:27.964015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 09:36:27.964018 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 09:36:27.964022 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1001 09:36:27.966112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.888763 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7aa5c938897ea621fc7c71c86e6a1dd2acca273e13afe71e9a27fc81f34cf2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.906221 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c68b374f72c2a689eb09480e7ae3e88691daf0cb380a52f655eb1c651a66a281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.908231 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.908303 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.908315 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.908331 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.908342 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:38Z","lastTransitionTime":"2025-10-01T09:36:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.930491 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72de5132325b6af956d84387ae27126362df626f66d135a50e420fc28bc0edc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72de5132325b6af956d84387ae27126362df626f66d135a50e420fc28bc0edc0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:36:37Z\\\",\\\"message\\\":\\\"olicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 09:36:37.858749 6046 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1001 09:36:37.858771 6046 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1001 09:36:37.858804 6046 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1001 09:36:37.858820 6046 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1001 09:36:37.859255 6046 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1001 09:36:37.859348 6046 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1001 09:36:37.859386 6046 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1001 09:36:37.859478 6046 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1001 09:36:37.859503 6046 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1001 09:36:37.859536 6046 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1001 09:36:37.859537 6046 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1001 09:36:37.859578 6046 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1001 09:36:37.859611 6046 factory.go:656] Stopping watch factory\\\\nI1001 09:36:37.859642 6046 ovnkube.go:599] Stopped ovnkube\\\\nI1001 09:36:37.859658 6046 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1001 09:36:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.943371 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc37698-af76-4818-b216-ffe35af67929\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1bb16edb25b63e957cc0bceb04a9fa9244e68881cb45be7c7d627c5faa74232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://995030421e0cbfeb7baa8ccb574d269dd178f8185bcaa9000fe5e7fc221b1188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d4c78dd607b81a28bcfe0ffa69c02628cf06aff5fd6a574069639c5bac044a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://144123d273f3f1af4fb857d731a16d9891b097c98229a035eb132f360fa8b2f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.963191 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72701c6061105b7bb81545b5f0b035b220b22b2f8009e82809a037870c625685\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.978709 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:38 crc kubenswrapper[4787]: I1001 09:36:38.996159 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.011839 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.011871 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.011882 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.011897 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.011908 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:39Z","lastTransitionTime":"2025-10-01T09:36:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.012111 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:39Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.023685 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebc8a6e2b9c6104b0f773e99d68a62e1b2bf0adc05e6f291cf91e7a8cf1bc121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:39Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.113917 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.113948 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.113956 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.113969 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.113977 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:39Z","lastTransitionTime":"2025-10-01T09:36:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.216689 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.216726 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.216735 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.216748 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.216757 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:39Z","lastTransitionTime":"2025-10-01T09:36:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.319773 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.319819 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.319835 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.319857 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.319875 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:39Z","lastTransitionTime":"2025-10-01T09:36:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.423111 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.423167 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.423180 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.423198 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.423213 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:39Z","lastTransitionTime":"2025-10-01T09:36:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.523509 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:36:39 crc kubenswrapper[4787]: E1001 09:36:39.523668 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.525176 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.525238 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.525268 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.525284 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.525294 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:39Z","lastTransitionTime":"2025-10-01T09:36:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.628145 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.628195 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.628205 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.628220 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.628231 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:39Z","lastTransitionTime":"2025-10-01T09:36:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.730846 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.730899 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.730915 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.730937 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.730951 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:39Z","lastTransitionTime":"2025-10-01T09:36:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.752746 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dmsxm_eadf2545-886e-4642-b909-704c6cd4134e/ovnkube-controller/0.log" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.757943 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" event={"ID":"eadf2545-886e-4642-b909-704c6cd4134e","Type":"ContainerStarted","Data":"899d7fc36a9f7a7335856902f4e1c915230f9b7e150d4139aeaab0652d74074c"} Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.758114 4787 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.780795 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:39Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.795625 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4990da342a20f6ab1e97774294bd6856c49ca17de21d893a638c13c8e727dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:39Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.811410 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b093e96a244fdec4c5cdce62650d4da31c2ec46207a9c08bbd0a28bf286abb52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4664cb1dde25ada88989e81f0319aedc6dbd363f608a50623efaa3eeb8d00763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:39Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.823814 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d95f8b1aac879d1746d8a8af7e1d5e672b36c31d5b95c23fee70b237bc3716e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:39Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.833891 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.834202 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.834332 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.834424 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.834514 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:39Z","lastTransitionTime":"2025-10-01T09:36:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.836374 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7aa5c938897ea621fc7c71c86e6a1dd2acca273e13afe71e9a27fc81f34cf2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:39Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.858039 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c68b374f72c2a689eb09480e7ae3e88691daf0cb380a52f655eb1c651a66a281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:39Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.880436 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://899d7fc36a9f7a7335856902f4e1c915230f9b7e150d4139aeaab0652d74074c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72de5132325b6af956d84387ae27126362df626f66d135a50e420fc28bc0edc0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:36:37Z\\\",\\\"message\\\":\\\"olicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 09:36:37.858749 6046 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1001 09:36:37.858771 6046 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1001 09:36:37.858804 6046 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1001 09:36:37.858820 6046 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1001 09:36:37.859255 6046 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1001 09:36:37.859348 6046 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1001 09:36:37.859386 6046 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1001 09:36:37.859478 6046 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1001 09:36:37.859503 6046 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1001 09:36:37.859536 6046 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1001 09:36:37.859537 6046 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1001 09:36:37.859578 6046 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1001 09:36:37.859611 6046 factory.go:656] Stopping watch factory\\\\nI1001 09:36:37.859642 6046 ovnkube.go:599] Stopped ovnkube\\\\nI1001 09:36:37.859658 6046 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1001 09:36:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:39Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.895454 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0814bed7-1117-4dff-97ef-a3cd1240e615\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c91542e7254c8289b734b3ffa8e8226f2445c915c264e09c179b99604abad1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:36:27Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1001 09:36:22.478145 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 09:36:22.478910 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3375976971/tls.crt::/tmp/serving-cert-3375976971/tls.key\\\\\\\"\\\\nI1001 09:36:27.943129 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 09:36:27.955126 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 09:36:27.955153 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 09:36:27.955179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 09:36:27.955185 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 09:36:27.963955 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1001 09:36:27.963979 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1001 09:36:27.963986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964008 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 09:36:27.964015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 09:36:27.964018 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 09:36:27.964022 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1001 09:36:27.966112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:39Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.910048 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc37698-af76-4818-b216-ffe35af67929\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1bb16edb25b63e957cc0bceb04a9fa9244e68881cb45be7c7d627c5faa74232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://995030421e0cbfeb7baa8ccb574d269dd178f8185bcaa9000fe5e7fc221b1188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d4c78dd607b81a28bcfe0ffa69c02628cf06aff5fd6a574069639c5bac044a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://144123d273f3f1af4fb857d731a16d9891b097c98229a035eb132f360fa8b2f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:39Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.926913 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72701c6061105b7bb81545b5f0b035b220b22b2f8009e82809a037870c625685\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:39Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.936801 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.936858 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.936869 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.936887 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.936901 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:39Z","lastTransitionTime":"2025-10-01T09:36:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.944619 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:39Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.962228 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:39Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.973895 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebc8a6e2b9c6104b0f773e99d68a62e1b2bf0adc05e6f291cf91e7a8cf1bc121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:39Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:39 crc kubenswrapper[4787]: I1001 09:36:39.989464 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:39Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.040265 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.040861 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.041057 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.041400 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.041642 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:40Z","lastTransitionTime":"2025-10-01T09:36:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.145987 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.146041 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.146057 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.146115 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.146130 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:40Z","lastTransitionTime":"2025-10-01T09:36:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.248836 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.248917 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.248936 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.248967 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.248987 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:40Z","lastTransitionTime":"2025-10-01T09:36:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.352260 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.352323 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.352340 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.352363 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.352380 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:40Z","lastTransitionTime":"2025-10-01T09:36:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.454576 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.455288 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.455358 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.455391 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.455408 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:40Z","lastTransitionTime":"2025-10-01T09:36:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.523311 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.523371 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:36:40 crc kubenswrapper[4787]: E1001 09:36:40.523436 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:36:40 crc kubenswrapper[4787]: E1001 09:36:40.523487 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.557964 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.558234 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.558382 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.558478 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.558581 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:40Z","lastTransitionTime":"2025-10-01T09:36:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.622781 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.660997 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.661338 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.661710 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.662047 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.662407 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:40Z","lastTransitionTime":"2025-10-01T09:36:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.762942 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dmsxm_eadf2545-886e-4642-b909-704c6cd4134e/ovnkube-controller/1.log" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.763755 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dmsxm_eadf2545-886e-4642-b909-704c6cd4134e/ovnkube-controller/0.log" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.764852 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.764880 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.764890 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.764907 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.764919 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:40Z","lastTransitionTime":"2025-10-01T09:36:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.768486 4787 generic.go:334] "Generic (PLEG): container finished" podID="eadf2545-886e-4642-b909-704c6cd4134e" containerID="899d7fc36a9f7a7335856902f4e1c915230f9b7e150d4139aeaab0652d74074c" exitCode=1 Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.768562 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" event={"ID":"eadf2545-886e-4642-b909-704c6cd4134e","Type":"ContainerDied","Data":"899d7fc36a9f7a7335856902f4e1c915230f9b7e150d4139aeaab0652d74074c"} Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.768692 4787 scope.go:117] "RemoveContainer" containerID="72de5132325b6af956d84387ae27126362df626f66d135a50e420fc28bc0edc0" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.769825 4787 scope.go:117] "RemoveContainer" containerID="899d7fc36a9f7a7335856902f4e1c915230f9b7e150d4139aeaab0652d74074c" Oct 01 09:36:40 crc kubenswrapper[4787]: E1001 09:36:40.770189 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dmsxm_openshift-ovn-kubernetes(eadf2545-886e-4642-b909-704c6cd4134e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" podUID="eadf2545-886e-4642-b909-704c6cd4134e" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.782854 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebc8a6e2b9c6104b0f773e99d68a62e1b2bf0adc05e6f291cf91e7a8cf1bc121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:40Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.805276 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:40Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.818222 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:40Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.830753 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4990da342a20f6ab1e97774294bd6856c49ca17de21d893a638c13c8e727dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:40Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.841793 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b093e96a244fdec4c5cdce62650d4da31c2ec46207a9c08bbd0a28bf286abb52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4664cb1dde25ada88989e81f0319aedc6dbd363f608a50623efaa3eeb8d00763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:40Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.854829 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d95f8b1aac879d1746d8a8af7e1d5e672b36c31d5b95c23fee70b237bc3716e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:40Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.868427 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.868468 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.868479 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.868495 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.868507 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:40Z","lastTransitionTime":"2025-10-01T09:36:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.869247 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7aa5c938897ea621fc7c71c86e6a1dd2acca273e13afe71e9a27fc81f34cf2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:40Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.886687 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c68b374f72c2a689eb09480e7ae3e88691daf0cb380a52f655eb1c651a66a281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:40Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.892528 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m"] Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.892968 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.894464 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.895440 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.909172 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://899d7fc36a9f7a7335856902f4e1c915230f9b7e150d4139aeaab0652d74074c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72de5132325b6af956d84387ae27126362df626f66d135a50e420fc28bc0edc0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:36:37Z\\\",\\\"message\\\":\\\"olicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 09:36:37.858749 6046 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1001 09:36:37.858771 6046 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1001 09:36:37.858804 6046 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1001 09:36:37.858820 6046 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1001 09:36:37.859255 6046 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1001 09:36:37.859348 6046 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1001 09:36:37.859386 6046 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1001 09:36:37.859478 6046 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1001 09:36:37.859503 6046 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1001 09:36:37.859536 6046 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1001 09:36:37.859537 6046 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1001 09:36:37.859578 6046 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1001 09:36:37.859611 6046 factory.go:656] Stopping watch factory\\\\nI1001 09:36:37.859642 6046 ovnkube.go:599] Stopped ovnkube\\\\nI1001 09:36:37.859658 6046 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1001 09:36:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899d7fc36a9f7a7335856902f4e1c915230f9b7e150d4139aeaab0652d74074c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:36:39Z\\\",\\\"message\\\":\\\"r/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1001 09:36:39.835210 6191 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nF1001 09:36:39.835211 6191 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:39Z is after 2025-08-24T17:21:41Z]\\\\nI1001 09:36:39.835216 6191 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1001 09:36:39.835221 6191 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1001 09:36:39.835225 6191 ovn.go:134] Ensuring zon\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:40Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.923230 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0814bed7-1117-4dff-97ef-a3cd1240e615\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c91542e7254c8289b734b3ffa8e8226f2445c915c264e09c179b99604abad1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:36:27Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1001 09:36:22.478145 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 09:36:22.478910 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3375976971/tls.crt::/tmp/serving-cert-3375976971/tls.key\\\\\\\"\\\\nI1001 09:36:27.943129 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 09:36:27.955126 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 09:36:27.955153 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 09:36:27.955179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 09:36:27.955185 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 09:36:27.963955 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1001 09:36:27.963979 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1001 09:36:27.963986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964008 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 09:36:27.964015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 09:36:27.964018 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 09:36:27.964022 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1001 09:36:27.966112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:40Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.935474 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc37698-af76-4818-b216-ffe35af67929\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1bb16edb25b63e957cc0bceb04a9fa9244e68881cb45be7c7d627c5faa74232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://995030421e0cbfeb7baa8ccb574d269dd178f8185bcaa9000fe5e7fc221b1188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d4c78dd607b81a28bcfe0ffa69c02628cf06aff5fd6a574069639c5bac044a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://144123d273f3f1af4fb857d731a16d9891b097c98229a035eb132f360fa8b2f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:40Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.948423 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72701c6061105b7bb81545b5f0b035b220b22b2f8009e82809a037870c625685\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:40Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.961673 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:40Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.970836 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.970864 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.970873 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.970886 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.970897 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:40Z","lastTransitionTime":"2025-10-01T09:36:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.975575 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:40Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:40 crc kubenswrapper[4787]: I1001 09:36:40.991433 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:40Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.004540 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4990da342a20f6ab1e97774294bd6856c49ca17de21d893a638c13c8e727dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.016583 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b093e96a244fdec4c5cdce62650d4da31c2ec46207a9c08bbd0a28bf286abb52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4664cb1dde25ada88989e81f0319aedc6dbd363f608a50623efaa3eeb8d00763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.026940 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d95f8b1aac879d1746d8a8af7e1d5e672b36c31d5b95c23fee70b237bc3716e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.038741 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0814bed7-1117-4dff-97ef-a3cd1240e615\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c91542e7254c8289b734b3ffa8e8226f2445c915c264e09c179b99604abad1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:36:27Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1001 09:36:22.478145 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 09:36:22.478910 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3375976971/tls.crt::/tmp/serving-cert-3375976971/tls.key\\\\\\\"\\\\nI1001 09:36:27.943129 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 09:36:27.955126 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 09:36:27.955153 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 09:36:27.955179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 09:36:27.955185 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 09:36:27.963955 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1001 09:36:27.963979 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1001 09:36:27.963986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964008 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 09:36:27.964015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 09:36:27.964018 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 09:36:27.964022 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1001 09:36:27.966112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.049209 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7aa5c938897ea621fc7c71c86e6a1dd2acca273e13afe71e9a27fc81f34cf2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.062502 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c68b374f72c2a689eb09480e7ae3e88691daf0cb380a52f655eb1c651a66a281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.062758 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/049825a6-0cec-4adf-8b77-4889bdc4d56f-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-84q9m\" (UID: \"049825a6-0cec-4adf-8b77-4889bdc4d56f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.063044 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/049825a6-0cec-4adf-8b77-4889bdc4d56f-env-overrides\") pod \"ovnkube-control-plane-749d76644c-84q9m\" (UID: \"049825a6-0cec-4adf-8b77-4889bdc4d56f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.063121 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqlsn\" (UniqueName: \"kubernetes.io/projected/049825a6-0cec-4adf-8b77-4889bdc4d56f-kube-api-access-sqlsn\") pod \"ovnkube-control-plane-749d76644c-84q9m\" (UID: \"049825a6-0cec-4adf-8b77-4889bdc4d56f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.063154 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/049825a6-0cec-4adf-8b77-4889bdc4d56f-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-84q9m\" (UID: \"049825a6-0cec-4adf-8b77-4889bdc4d56f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.073498 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.073531 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.073540 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.073553 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.073570 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:41Z","lastTransitionTime":"2025-10-01T09:36:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.079277 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://899d7fc36a9f7a7335856902f4e1c915230f9b7e150d4139aeaab0652d74074c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72de5132325b6af956d84387ae27126362df626f66d135a50e420fc28bc0edc0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:36:37Z\\\",\\\"message\\\":\\\"olicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 09:36:37.858749 6046 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1001 09:36:37.858771 6046 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1001 09:36:37.858804 6046 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1001 09:36:37.858820 6046 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1001 09:36:37.859255 6046 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1001 09:36:37.859348 6046 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1001 09:36:37.859386 6046 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1001 09:36:37.859478 6046 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1001 09:36:37.859503 6046 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1001 09:36:37.859536 6046 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1001 09:36:37.859537 6046 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1001 09:36:37.859578 6046 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1001 09:36:37.859611 6046 factory.go:656] Stopping watch factory\\\\nI1001 09:36:37.859642 6046 ovnkube.go:599] Stopped ovnkube\\\\nI1001 09:36:37.859658 6046 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1001 09:36:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899d7fc36a9f7a7335856902f4e1c915230f9b7e150d4139aeaab0652d74074c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:36:39Z\\\",\\\"message\\\":\\\"r/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1001 09:36:39.835210 6191 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nF1001 09:36:39.835211 6191 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:39Z is after 2025-08-24T17:21:41Z]\\\\nI1001 09:36:39.835216 6191 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1001 09:36:39.835221 6191 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1001 09:36:39.835225 6191 ovn.go:134] Ensuring zon\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.089734 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc37698-af76-4818-b216-ffe35af67929\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1bb16edb25b63e957cc0bceb04a9fa9244e68881cb45be7c7d627c5faa74232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://995030421e0cbfeb7baa8ccb574d269dd178f8185bcaa9000fe5e7fc221b1188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d4c78dd607b81a28bcfe0ffa69c02628cf06aff5fd6a574069639c5bac044a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://144123d273f3f1af4fb857d731a16d9891b097c98229a035eb132f360fa8b2f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.100782 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72701c6061105b7bb81545b5f0b035b220b22b2f8009e82809a037870c625685\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.110986 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.123729 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.133727 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.144378 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebc8a6e2b9c6104b0f773e99d68a62e1b2bf0adc05e6f291cf91e7a8cf1bc121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.157120 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"049825a6-0cec-4adf-8b77-4889bdc4d56f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-84q9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.164480 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/049825a6-0cec-4adf-8b77-4889bdc4d56f-env-overrides\") pod \"ovnkube-control-plane-749d76644c-84q9m\" (UID: \"049825a6-0cec-4adf-8b77-4889bdc4d56f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.164530 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqlsn\" (UniqueName: \"kubernetes.io/projected/049825a6-0cec-4adf-8b77-4889bdc4d56f-kube-api-access-sqlsn\") pod \"ovnkube-control-plane-749d76644c-84q9m\" (UID: \"049825a6-0cec-4adf-8b77-4889bdc4d56f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.164559 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/049825a6-0cec-4adf-8b77-4889bdc4d56f-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-84q9m\" (UID: \"049825a6-0cec-4adf-8b77-4889bdc4d56f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.164582 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/049825a6-0cec-4adf-8b77-4889bdc4d56f-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-84q9m\" (UID: \"049825a6-0cec-4adf-8b77-4889bdc4d56f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.165171 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/049825a6-0cec-4adf-8b77-4889bdc4d56f-env-overrides\") pod \"ovnkube-control-plane-749d76644c-84q9m\" (UID: \"049825a6-0cec-4adf-8b77-4889bdc4d56f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.165261 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/049825a6-0cec-4adf-8b77-4889bdc4d56f-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-84q9m\" (UID: \"049825a6-0cec-4adf-8b77-4889bdc4d56f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.169383 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/049825a6-0cec-4adf-8b77-4889bdc4d56f-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-84q9m\" (UID: \"049825a6-0cec-4adf-8b77-4889bdc4d56f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.176322 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.176352 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.176362 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.176377 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.176387 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:41Z","lastTransitionTime":"2025-10-01T09:36:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.179546 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqlsn\" (UniqueName: \"kubernetes.io/projected/049825a6-0cec-4adf-8b77-4889bdc4d56f-kube-api-access-sqlsn\") pod \"ovnkube-control-plane-749d76644c-84q9m\" (UID: \"049825a6-0cec-4adf-8b77-4889bdc4d56f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.207870 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m" Oct 01 09:36:41 crc kubenswrapper[4787]: W1001 09:36:41.219862 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod049825a6_0cec_4adf_8b77_4889bdc4d56f.slice/crio-b8d3ad6269e00f1da1bf0b6ff8d2afdebcc1c6836b7a89066039a4e20551360e WatchSource:0}: Error finding container b8d3ad6269e00f1da1bf0b6ff8d2afdebcc1c6836b7a89066039a4e20551360e: Status 404 returned error can't find the container with id b8d3ad6269e00f1da1bf0b6ff8d2afdebcc1c6836b7a89066039a4e20551360e Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.278640 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.278684 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.278697 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.278714 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.278726 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:41Z","lastTransitionTime":"2025-10-01T09:36:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.380583 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.380617 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.380631 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.380647 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.380659 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:41Z","lastTransitionTime":"2025-10-01T09:36:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.482622 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.482664 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.482676 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.482692 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.482704 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:41Z","lastTransitionTime":"2025-10-01T09:36:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.523112 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:36:41 crc kubenswrapper[4787]: E1001 09:36:41.523260 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.584561 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.584606 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.584618 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.584636 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.584648 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:41Z","lastTransitionTime":"2025-10-01T09:36:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.687059 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.687124 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.687133 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.687150 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.687161 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:41Z","lastTransitionTime":"2025-10-01T09:36:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.777264 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m" event={"ID":"049825a6-0cec-4adf-8b77-4889bdc4d56f","Type":"ContainerStarted","Data":"8bc8192c84986df76c9a45deba61aaaa1b0a019fce0000e1c9916c4fe31d1ba1"} Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.777325 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m" event={"ID":"049825a6-0cec-4adf-8b77-4889bdc4d56f","Type":"ContainerStarted","Data":"18044055eb1f6cfd7cb05f330314231b1c563c34a51ba5ed79073988eb563cc3"} Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.777342 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m" event={"ID":"049825a6-0cec-4adf-8b77-4889bdc4d56f","Type":"ContainerStarted","Data":"b8d3ad6269e00f1da1bf0b6ff8d2afdebcc1c6836b7a89066039a4e20551360e"} Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.778784 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dmsxm_eadf2545-886e-4642-b909-704c6cd4134e/ovnkube-controller/1.log" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.782749 4787 scope.go:117] "RemoveContainer" containerID="899d7fc36a9f7a7335856902f4e1c915230f9b7e150d4139aeaab0652d74074c" Oct 01 09:36:41 crc kubenswrapper[4787]: E1001 09:36:41.782954 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dmsxm_openshift-ovn-kubernetes(eadf2545-886e-4642-b909-704c6cd4134e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" podUID="eadf2545-886e-4642-b909-704c6cd4134e" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.789447 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.789487 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.789498 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.789513 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.789524 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:41Z","lastTransitionTime":"2025-10-01T09:36:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.790897 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.801455 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4990da342a20f6ab1e97774294bd6856c49ca17de21d893a638c13c8e727dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.813065 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b093e96a244fdec4c5cdce62650d4da31c2ec46207a9c08bbd0a28bf286abb52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4664cb1dde25ada88989e81f0319aedc6dbd363f608a50623efaa3eeb8d00763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.822584 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d95f8b1aac879d1746d8a8af7e1d5e672b36c31d5b95c23fee70b237bc3716e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.834196 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0814bed7-1117-4dff-97ef-a3cd1240e615\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c91542e7254c8289b734b3ffa8e8226f2445c915c264e09c179b99604abad1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:36:27Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1001 09:36:22.478145 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 09:36:22.478910 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3375976971/tls.crt::/tmp/serving-cert-3375976971/tls.key\\\\\\\"\\\\nI1001 09:36:27.943129 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 09:36:27.955126 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 09:36:27.955153 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 09:36:27.955179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 09:36:27.955185 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 09:36:27.963955 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1001 09:36:27.963979 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1001 09:36:27.963986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964008 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 09:36:27.964015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 09:36:27.964018 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 09:36:27.964022 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1001 09:36:27.966112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.843035 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7aa5c938897ea621fc7c71c86e6a1dd2acca273e13afe71e9a27fc81f34cf2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.856660 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c68b374f72c2a689eb09480e7ae3e88691daf0cb380a52f655eb1c651a66a281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.877789 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://899d7fc36a9f7a7335856902f4e1c915230f9b7e150d4139aeaab0652d74074c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72de5132325b6af956d84387ae27126362df626f66d135a50e420fc28bc0edc0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:36:37Z\\\",\\\"message\\\":\\\"olicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1001 09:36:37.858749 6046 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1001 09:36:37.858771 6046 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1001 09:36:37.858804 6046 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1001 09:36:37.858820 6046 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1001 09:36:37.859255 6046 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1001 09:36:37.859348 6046 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1001 09:36:37.859386 6046 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1001 09:36:37.859478 6046 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1001 09:36:37.859503 6046 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1001 09:36:37.859536 6046 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1001 09:36:37.859537 6046 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1001 09:36:37.859578 6046 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1001 09:36:37.859611 6046 factory.go:656] Stopping watch factory\\\\nI1001 09:36:37.859642 6046 ovnkube.go:599] Stopped ovnkube\\\\nI1001 09:36:37.859658 6046 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1001 09:36:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899d7fc36a9f7a7335856902f4e1c915230f9b7e150d4139aeaab0652d74074c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:36:39Z\\\",\\\"message\\\":\\\"r/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1001 09:36:39.835210 6191 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nF1001 09:36:39.835211 6191 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:39Z is after 2025-08-24T17:21:41Z]\\\\nI1001 09:36:39.835216 6191 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1001 09:36:39.835221 6191 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1001 09:36:39.835225 6191 ovn.go:134] Ensuring zon\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.891602 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.891647 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.891665 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.891688 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.891703 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:41Z","lastTransitionTime":"2025-10-01T09:36:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.893445 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc37698-af76-4818-b216-ffe35af67929\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1bb16edb25b63e957cc0bceb04a9fa9244e68881cb45be7c7d627c5faa74232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://995030421e0cbfeb7baa8ccb574d269dd178f8185bcaa9000fe5e7fc221b1188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d4c78dd607b81a28bcfe0ffa69c02628cf06aff5fd6a574069639c5bac044a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://144123d273f3f1af4fb857d731a16d9891b097c98229a035eb132f360fa8b2f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.909863 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72701c6061105b7bb81545b5f0b035b220b22b2f8009e82809a037870c625685\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.922880 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.934357 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.944870 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.953348 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebc8a6e2b9c6104b0f773e99d68a62e1b2bf0adc05e6f291cf91e7a8cf1bc121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.962291 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"049825a6-0cec-4adf-8b77-4889bdc4d56f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18044055eb1f6cfd7cb05f330314231b1c563c34a51ba5ed79073988eb563cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc8192c84986df76c9a45deba61aaaa1b0a019fce0000e1c9916c4fe31d1ba1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-84q9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.972116 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.980659 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebc8a6e2b9c6104b0f773e99d68a62e1b2bf0adc05e6f291cf91e7a8cf1bc121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.990316 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"049825a6-0cec-4adf-8b77-4889bdc4d56f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18044055eb1f6cfd7cb05f330314231b1c563c34a51ba5ed79073988eb563cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc8192c84986df76c9a45deba61aaaa1b0a019fce0000e1c9916c4fe31d1ba1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-84q9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:41Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.993711 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.993745 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.993754 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.993767 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:41 crc kubenswrapper[4787]: I1001 09:36:41.993777 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:41Z","lastTransitionTime":"2025-10-01T09:36:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.002863 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b093e96a244fdec4c5cdce62650d4da31c2ec46207a9c08bbd0a28bf286abb52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4664cb1dde25ada88989e81f0319aedc6dbd363f608a50623efaa3eeb8d00763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.012111 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d95f8b1aac879d1746d8a8af7e1d5e672b36c31d5b95c23fee70b237bc3716e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.020625 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.031203 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4990da342a20f6ab1e97774294bd6856c49ca17de21d893a638c13c8e727dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.042112 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0814bed7-1117-4dff-97ef-a3cd1240e615\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c91542e7254c8289b734b3ffa8e8226f2445c915c264e09c179b99604abad1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:36:27Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1001 09:36:22.478145 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 09:36:22.478910 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3375976971/tls.crt::/tmp/serving-cert-3375976971/tls.key\\\\\\\"\\\\nI1001 09:36:27.943129 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 09:36:27.955126 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 09:36:27.955153 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 09:36:27.955179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 09:36:27.955185 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 09:36:27.963955 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1001 09:36:27.963979 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1001 09:36:27.963986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964008 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 09:36:27.964015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 09:36:27.964018 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 09:36:27.964022 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1001 09:36:27.966112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.050311 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7aa5c938897ea621fc7c71c86e6a1dd2acca273e13afe71e9a27fc81f34cf2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.062238 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c68b374f72c2a689eb09480e7ae3e88691daf0cb380a52f655eb1c651a66a281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.078003 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://899d7fc36a9f7a7335856902f4e1c915230f9b7e150d4139aeaab0652d74074c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899d7fc36a9f7a7335856902f4e1c915230f9b7e150d4139aeaab0652d74074c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:36:39Z\\\",\\\"message\\\":\\\"r/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1001 09:36:39.835210 6191 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nF1001 09:36:39.835211 6191 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:39Z is after 2025-08-24T17:21:41Z]\\\\nI1001 09:36:39.835216 6191 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1001 09:36:39.835221 6191 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1001 09:36:39.835225 6191 ovn.go:134] Ensuring zon\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dmsxm_openshift-ovn-kubernetes(eadf2545-886e-4642-b909-704c6cd4134e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.088304 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.095886 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.095925 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.095937 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.095953 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.095964 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:42Z","lastTransitionTime":"2025-10-01T09:36:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.102456 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc37698-af76-4818-b216-ffe35af67929\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1bb16edb25b63e957cc0bceb04a9fa9244e68881cb45be7c7d627c5faa74232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://995030421e0cbfeb7baa8ccb574d269dd178f8185bcaa9000fe5e7fc221b1188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d4c78dd607b81a28bcfe0ffa69c02628cf06aff5fd6a574069639c5bac044a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://144123d273f3f1af4fb857d731a16d9891b097c98229a035eb132f360fa8b2f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.117292 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72701c6061105b7bb81545b5f0b035b220b22b2f8009e82809a037870c625685\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.127469 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.198423 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.198474 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.198483 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.198496 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.198504 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:42Z","lastTransitionTime":"2025-10-01T09:36:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.300462 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.300706 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.300791 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.300871 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.300933 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:42Z","lastTransitionTime":"2025-10-01T09:36:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.368376 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-6zz6g"] Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.368788 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:36:42 crc kubenswrapper[4787]: E1001 09:36:42.368837 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.390299 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0814bed7-1117-4dff-97ef-a3cd1240e615\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c91542e7254c8289b734b3ffa8e8226f2445c915c264e09c179b99604abad1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:36:27Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1001 09:36:22.478145 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 09:36:22.478910 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3375976971/tls.crt::/tmp/serving-cert-3375976971/tls.key\\\\\\\"\\\\nI1001 09:36:27.943129 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 09:36:27.955126 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 09:36:27.955153 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 09:36:27.955179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 09:36:27.955185 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 09:36:27.963955 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1001 09:36:27.963979 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1001 09:36:27.963986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964008 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 09:36:27.964015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 09:36:27.964018 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 09:36:27.964022 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1001 09:36:27.966112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.403796 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.403850 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.403866 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.403891 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.403909 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:42Z","lastTransitionTime":"2025-10-01T09:36:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.404372 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7aa5c938897ea621fc7c71c86e6a1dd2acca273e13afe71e9a27fc81f34cf2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.421392 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c68b374f72c2a689eb09480e7ae3e88691daf0cb380a52f655eb1c651a66a281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.442178 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://899d7fc36a9f7a7335856902f4e1c915230f9b7e150d4139aeaab0652d74074c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899d7fc36a9f7a7335856902f4e1c915230f9b7e150d4139aeaab0652d74074c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:36:39Z\\\",\\\"message\\\":\\\"r/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1001 09:36:39.835210 6191 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nF1001 09:36:39.835211 6191 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:39Z is after 2025-08-24T17:21:41Z]\\\\nI1001 09:36:39.835216 6191 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1001 09:36:39.835221 6191 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1001 09:36:39.835225 6191 ovn.go:134] Ensuring zon\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dmsxm_openshift-ovn-kubernetes(eadf2545-886e-4642-b909-704c6cd4134e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.456404 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6zz6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb4a902-9efa-4088-a236-9a634fd085fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k5bxp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k5bxp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6zz6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.471408 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc37698-af76-4818-b216-ffe35af67929\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1bb16edb25b63e957cc0bceb04a9fa9244e68881cb45be7c7d627c5faa74232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://995030421e0cbfeb7baa8ccb574d269dd178f8185bcaa9000fe5e7fc221b1188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d4c78dd607b81a28bcfe0ffa69c02628cf06aff5fd6a574069639c5bac044a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://144123d273f3f1af4fb857d731a16d9891b097c98229a035eb132f360fa8b2f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.474956 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5bxp\" (UniqueName: \"kubernetes.io/projected/0eb4a902-9efa-4088-a236-9a634fd085fd-kube-api-access-k5bxp\") pod \"network-metrics-daemon-6zz6g\" (UID: \"0eb4a902-9efa-4088-a236-9a634fd085fd\") " pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.475170 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0eb4a902-9efa-4088-a236-9a634fd085fd-metrics-certs\") pod \"network-metrics-daemon-6zz6g\" (UID: \"0eb4a902-9efa-4088-a236-9a634fd085fd\") " pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.491393 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72701c6061105b7bb81545b5f0b035b220b22b2f8009e82809a037870c625685\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.504056 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.506273 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.506306 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.506317 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.506332 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.506343 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:42Z","lastTransitionTime":"2025-10-01T09:36:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.518755 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.522890 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.522936 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:36:42 crc kubenswrapper[4787]: E1001 09:36:42.522988 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:36:42 crc kubenswrapper[4787]: E1001 09:36:42.523060 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.533205 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.544155 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebc8a6e2b9c6104b0f773e99d68a62e1b2bf0adc05e6f291cf91e7a8cf1bc121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.555585 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"049825a6-0cec-4adf-8b77-4889bdc4d56f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18044055eb1f6cfd7cb05f330314231b1c563c34a51ba5ed79073988eb563cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc8192c84986df76c9a45deba61aaaa1b0a019fce0000e1c9916c4fe31d1ba1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-84q9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.567476 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.575642 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5bxp\" (UniqueName: \"kubernetes.io/projected/0eb4a902-9efa-4088-a236-9a634fd085fd-kube-api-access-k5bxp\") pod \"network-metrics-daemon-6zz6g\" (UID: \"0eb4a902-9efa-4088-a236-9a634fd085fd\") " pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.575691 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0eb4a902-9efa-4088-a236-9a634fd085fd-metrics-certs\") pod \"network-metrics-daemon-6zz6g\" (UID: \"0eb4a902-9efa-4088-a236-9a634fd085fd\") " pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:36:42 crc kubenswrapper[4787]: E1001 09:36:42.575788 4787 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 09:36:42 crc kubenswrapper[4787]: E1001 09:36:42.575849 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0eb4a902-9efa-4088-a236-9a634fd085fd-metrics-certs podName:0eb4a902-9efa-4088-a236-9a634fd085fd nodeName:}" failed. No retries permitted until 2025-10-01 09:36:43.075835229 +0000 UTC m=+35.190979386 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0eb4a902-9efa-4088-a236-9a634fd085fd-metrics-certs") pod "network-metrics-daemon-6zz6g" (UID: "0eb4a902-9efa-4088-a236-9a634fd085fd") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.579934 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4990da342a20f6ab1e97774294bd6856c49ca17de21d893a638c13c8e727dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.592325 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b093e96a244fdec4c5cdce62650d4da31c2ec46207a9c08bbd0a28bf286abb52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4664cb1dde25ada88989e81f0319aedc6dbd363f608a50623efaa3eeb8d00763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.597268 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5bxp\" (UniqueName: \"kubernetes.io/projected/0eb4a902-9efa-4088-a236-9a634fd085fd-kube-api-access-k5bxp\") pod \"network-metrics-daemon-6zz6g\" (UID: \"0eb4a902-9efa-4088-a236-9a634fd085fd\") " pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.607006 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d95f8b1aac879d1746d8a8af7e1d5e672b36c31d5b95c23fee70b237bc3716e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:42Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.609051 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.609121 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.609137 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.609162 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.609177 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:42Z","lastTransitionTime":"2025-10-01T09:36:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.711846 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.711913 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.711923 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.711943 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.711956 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:42Z","lastTransitionTime":"2025-10-01T09:36:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.814886 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.814922 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.814935 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.814954 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.814966 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:42Z","lastTransitionTime":"2025-10-01T09:36:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.917189 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.917247 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.917269 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.917297 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:42 crc kubenswrapper[4787]: I1001 09:36:42.917323 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:42Z","lastTransitionTime":"2025-10-01T09:36:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.019880 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.019932 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.019953 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.019976 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.019994 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:43Z","lastTransitionTime":"2025-10-01T09:36:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.080056 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0eb4a902-9efa-4088-a236-9a634fd085fd-metrics-certs\") pod \"network-metrics-daemon-6zz6g\" (UID: \"0eb4a902-9efa-4088-a236-9a634fd085fd\") " pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:36:43 crc kubenswrapper[4787]: E1001 09:36:43.080406 4787 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 09:36:43 crc kubenswrapper[4787]: E1001 09:36:43.080546 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0eb4a902-9efa-4088-a236-9a634fd085fd-metrics-certs podName:0eb4a902-9efa-4088-a236-9a634fd085fd nodeName:}" failed. No retries permitted until 2025-10-01 09:36:44.080512992 +0000 UTC m=+36.195657179 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0eb4a902-9efa-4088-a236-9a634fd085fd-metrics-certs") pod "network-metrics-daemon-6zz6g" (UID: "0eb4a902-9efa-4088-a236-9a634fd085fd") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.122418 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.122482 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.122501 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.122521 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.122533 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:43Z","lastTransitionTime":"2025-10-01T09:36:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.225930 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.226001 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.226015 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.226039 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.226054 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:43Z","lastTransitionTime":"2025-10-01T09:36:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.329625 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.329692 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.329712 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.329739 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.329758 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:43Z","lastTransitionTime":"2025-10-01T09:36:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.432713 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.433176 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.433186 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.433206 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.433220 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:43Z","lastTransitionTime":"2025-10-01T09:36:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.523413 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:36:43 crc kubenswrapper[4787]: E1001 09:36:43.523657 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.536566 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.536619 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.536630 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.536649 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.536662 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:43Z","lastTransitionTime":"2025-10-01T09:36:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.638832 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.639013 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.639161 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.639238 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.639307 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:43Z","lastTransitionTime":"2025-10-01T09:36:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.743218 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.743265 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.743278 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.743296 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.743306 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:43Z","lastTransitionTime":"2025-10-01T09:36:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.845845 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.845893 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.845904 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.845923 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.845938 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:43Z","lastTransitionTime":"2025-10-01T09:36:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.949615 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.949710 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.949727 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.949769 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:43 crc kubenswrapper[4787]: I1001 09:36:43.949789 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:43Z","lastTransitionTime":"2025-10-01T09:36:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.053247 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.053299 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.053309 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.053329 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.053341 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:44Z","lastTransitionTime":"2025-10-01T09:36:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.098010 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0eb4a902-9efa-4088-a236-9a634fd085fd-metrics-certs\") pod \"network-metrics-daemon-6zz6g\" (UID: \"0eb4a902-9efa-4088-a236-9a634fd085fd\") " pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:36:44 crc kubenswrapper[4787]: E1001 09:36:44.098252 4787 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 09:36:44 crc kubenswrapper[4787]: E1001 09:36:44.098361 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0eb4a902-9efa-4088-a236-9a634fd085fd-metrics-certs podName:0eb4a902-9efa-4088-a236-9a634fd085fd nodeName:}" failed. No retries permitted until 2025-10-01 09:36:46.098329518 +0000 UTC m=+38.213473895 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0eb4a902-9efa-4088-a236-9a634fd085fd-metrics-certs") pod "network-metrics-daemon-6zz6g" (UID: "0eb4a902-9efa-4088-a236-9a634fd085fd") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.156717 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.156770 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.156778 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.156793 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.156806 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:44Z","lastTransitionTime":"2025-10-01T09:36:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.259997 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.260062 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.260095 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.260110 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.260119 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:44Z","lastTransitionTime":"2025-10-01T09:36:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.301156 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.301330 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:36:44 crc kubenswrapper[4787]: E1001 09:36:44.301370 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:37:00.30133765 +0000 UTC m=+52.416481807 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:36:44 crc kubenswrapper[4787]: E1001 09:36:44.301525 4787 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.301532 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:36:44 crc kubenswrapper[4787]: E1001 09:36:44.301626 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 09:37:00.301598536 +0000 UTC m=+52.416742723 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 09:36:44 crc kubenswrapper[4787]: E1001 09:36:44.301653 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 09:36:44 crc kubenswrapper[4787]: E1001 09:36:44.301668 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 09:36:44 crc kubenswrapper[4787]: E1001 09:36:44.301681 4787 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:36:44 crc kubenswrapper[4787]: E1001 09:36:44.301716 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-01 09:37:00.301709259 +0000 UTC m=+52.416853416 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.301719 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.301779 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:36:44 crc kubenswrapper[4787]: E1001 09:36:44.301916 4787 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 09:36:44 crc kubenswrapper[4787]: E1001 09:36:44.301957 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 09:36:44 crc kubenswrapper[4787]: E1001 09:36:44.302107 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 09:36:44 crc kubenswrapper[4787]: E1001 09:36:44.302121 4787 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:36:44 crc kubenswrapper[4787]: E1001 09:36:44.302140 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 09:37:00.302070227 +0000 UTC m=+52.417214414 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 09:36:44 crc kubenswrapper[4787]: E1001 09:36:44.302177 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-01 09:37:00.302157221 +0000 UTC m=+52.417301458 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.362031 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.362105 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.362116 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.362130 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.362141 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:44Z","lastTransitionTime":"2025-10-01T09:36:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.463911 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.463954 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.463967 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.464032 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.464045 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:44Z","lastTransitionTime":"2025-10-01T09:36:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.522949 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.523032 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:36:44 crc kubenswrapper[4787]: E1001 09:36:44.523261 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:36:44 crc kubenswrapper[4787]: E1001 09:36:44.523052 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.523449 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:36:44 crc kubenswrapper[4787]: E1001 09:36:44.523581 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.566551 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.566582 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.566592 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.566605 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.566616 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:44Z","lastTransitionTime":"2025-10-01T09:36:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.602410 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.602439 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.602450 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.602488 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.602502 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:44Z","lastTransitionTime":"2025-10-01T09:36:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:44 crc kubenswrapper[4787]: E1001 09:36:44.614781 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:44Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.617815 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.617869 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.617880 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.617895 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.617905 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:44Z","lastTransitionTime":"2025-10-01T09:36:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:44 crc kubenswrapper[4787]: E1001 09:36:44.634587 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:44Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.638258 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.638289 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.638299 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.638319 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.638330 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:44Z","lastTransitionTime":"2025-10-01T09:36:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:44 crc kubenswrapper[4787]: E1001 09:36:44.653901 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:44Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.657727 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.657759 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.657767 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.657783 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.657793 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:44Z","lastTransitionTime":"2025-10-01T09:36:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:44 crc kubenswrapper[4787]: E1001 09:36:44.673457 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:44Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.677440 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.677470 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.677479 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.677492 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.677501 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:44Z","lastTransitionTime":"2025-10-01T09:36:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:44 crc kubenswrapper[4787]: E1001 09:36:44.688026 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:44Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:44 crc kubenswrapper[4787]: E1001 09:36:44.688169 4787 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.689551 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.689605 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.689618 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.689637 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.689649 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:44Z","lastTransitionTime":"2025-10-01T09:36:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.791448 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.791516 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.791533 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.791546 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.791554 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:44Z","lastTransitionTime":"2025-10-01T09:36:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.894394 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.894450 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.894461 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.894481 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.894492 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:44Z","lastTransitionTime":"2025-10-01T09:36:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.999675 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.999726 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:44 crc kubenswrapper[4787]: I1001 09:36:44.999742 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:44.999766 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:44.999783 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:44Z","lastTransitionTime":"2025-10-01T09:36:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.102515 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.102596 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.102612 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.102630 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.102640 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:45Z","lastTransitionTime":"2025-10-01T09:36:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.205617 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.205649 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.205658 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.205673 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.205684 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:45Z","lastTransitionTime":"2025-10-01T09:36:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.309140 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.309205 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.309235 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.309256 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.309271 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:45Z","lastTransitionTime":"2025-10-01T09:36:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.414456 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.414531 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.414548 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.415132 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.415200 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:45Z","lastTransitionTime":"2025-10-01T09:36:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.517836 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.517904 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.517922 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.518346 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.518406 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:45Z","lastTransitionTime":"2025-10-01T09:36:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.523008 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:36:45 crc kubenswrapper[4787]: E1001 09:36:45.523110 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.621615 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.621690 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.621710 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.621740 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.621759 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:45Z","lastTransitionTime":"2025-10-01T09:36:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.725934 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.726008 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.726031 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.726061 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.726112 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:45Z","lastTransitionTime":"2025-10-01T09:36:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.827840 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.827877 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.827884 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.827897 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.827909 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:45Z","lastTransitionTime":"2025-10-01T09:36:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.930687 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.930727 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.930738 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.930753 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:45 crc kubenswrapper[4787]: I1001 09:36:45.930765 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:45Z","lastTransitionTime":"2025-10-01T09:36:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.033092 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.033374 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.033453 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.033560 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.033648 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:46Z","lastTransitionTime":"2025-10-01T09:36:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.117341 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0eb4a902-9efa-4088-a236-9a634fd085fd-metrics-certs\") pod \"network-metrics-daemon-6zz6g\" (UID: \"0eb4a902-9efa-4088-a236-9a634fd085fd\") " pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:36:46 crc kubenswrapper[4787]: E1001 09:36:46.117539 4787 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 09:36:46 crc kubenswrapper[4787]: E1001 09:36:46.117708 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0eb4a902-9efa-4088-a236-9a634fd085fd-metrics-certs podName:0eb4a902-9efa-4088-a236-9a634fd085fd nodeName:}" failed. No retries permitted until 2025-10-01 09:36:50.117694298 +0000 UTC m=+42.232838455 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0eb4a902-9efa-4088-a236-9a634fd085fd-metrics-certs") pod "network-metrics-daemon-6zz6g" (UID: "0eb4a902-9efa-4088-a236-9a634fd085fd") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.135666 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.135696 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.135704 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.135717 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.135725 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:46Z","lastTransitionTime":"2025-10-01T09:36:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.237983 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.238281 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.238438 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.238527 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.238612 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:46Z","lastTransitionTime":"2025-10-01T09:36:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.340852 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.341141 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.341218 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.341296 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.341364 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:46Z","lastTransitionTime":"2025-10-01T09:36:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.443883 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.443930 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.443941 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.443956 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.443966 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:46Z","lastTransitionTime":"2025-10-01T09:36:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.522791 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:36:46 crc kubenswrapper[4787]: E1001 09:36:46.522919 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.523151 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:36:46 crc kubenswrapper[4787]: E1001 09:36:46.523373 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.523396 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:36:46 crc kubenswrapper[4787]: E1001 09:36:46.523520 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.546625 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.546709 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.546754 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.546778 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.546824 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:46Z","lastTransitionTime":"2025-10-01T09:36:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.648960 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.649004 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.649014 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.649029 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.649039 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:46Z","lastTransitionTime":"2025-10-01T09:36:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.750956 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.751006 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.751016 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.751030 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.751040 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:46Z","lastTransitionTime":"2025-10-01T09:36:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.854357 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.854423 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.854437 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.854458 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.854473 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:46Z","lastTransitionTime":"2025-10-01T09:36:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.956921 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.957212 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.957280 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.957347 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:46 crc kubenswrapper[4787]: I1001 09:36:46.957406 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:46Z","lastTransitionTime":"2025-10-01T09:36:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.060576 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.060882 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.060960 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.061051 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.061153 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:47Z","lastTransitionTime":"2025-10-01T09:36:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.165326 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.165404 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.165423 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.165453 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.165475 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:47Z","lastTransitionTime":"2025-10-01T09:36:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.268235 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.268276 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.268284 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.268300 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.268311 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:47Z","lastTransitionTime":"2025-10-01T09:36:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.370904 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.370969 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.370981 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.370999 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.371012 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:47Z","lastTransitionTime":"2025-10-01T09:36:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.476440 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.476483 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.476494 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.476509 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.476520 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:47Z","lastTransitionTime":"2025-10-01T09:36:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.522973 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:36:47 crc kubenswrapper[4787]: E1001 09:36:47.523125 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.579895 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.579986 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.579997 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.580017 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.580028 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:47Z","lastTransitionTime":"2025-10-01T09:36:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.683211 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.683264 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.683278 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.683296 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.683310 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:47Z","lastTransitionTime":"2025-10-01T09:36:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.787339 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.787394 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.787411 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.787436 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.787450 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:47Z","lastTransitionTime":"2025-10-01T09:36:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.890572 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.891010 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.891139 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.891240 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.891340 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:47Z","lastTransitionTime":"2025-10-01T09:36:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.947065 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.963894 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:47Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.981523 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4990da342a20f6ab1e97774294bd6856c49ca17de21d893a638c13c8e727dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:47Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.993826 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.993909 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.993929 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.993961 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.993986 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:47Z","lastTransitionTime":"2025-10-01T09:36:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:47 crc kubenswrapper[4787]: I1001 09:36:47.994758 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b093e96a244fdec4c5cdce62650d4da31c2ec46207a9c08bbd0a28bf286abb52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4664cb1dde25ada88989e81f0319aedc6dbd363f608a50623efaa3eeb8d00763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:47Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.008113 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d95f8b1aac879d1746d8a8af7e1d5e672b36c31d5b95c23fee70b237bc3716e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.022567 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c68b374f72c2a689eb09480e7ae3e88691daf0cb380a52f655eb1c651a66a281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.039999 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://899d7fc36a9f7a7335856902f4e1c915230f9b7e150d4139aeaab0652d74074c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899d7fc36a9f7a7335856902f4e1c915230f9b7e150d4139aeaab0652d74074c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:36:39Z\\\",\\\"message\\\":\\\"r/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1001 09:36:39.835210 6191 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nF1001 09:36:39.835211 6191 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:39Z is after 2025-08-24T17:21:41Z]\\\\nI1001 09:36:39.835216 6191 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1001 09:36:39.835221 6191 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1001 09:36:39.835225 6191 ovn.go:134] Ensuring zon\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dmsxm_openshift-ovn-kubernetes(eadf2545-886e-4642-b909-704c6cd4134e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.050856 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6zz6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb4a902-9efa-4088-a236-9a634fd085fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k5bxp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k5bxp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6zz6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.065779 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0814bed7-1117-4dff-97ef-a3cd1240e615\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c91542e7254c8289b734b3ffa8e8226f2445c915c264e09c179b99604abad1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:36:27Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1001 09:36:22.478145 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 09:36:22.478910 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3375976971/tls.crt::/tmp/serving-cert-3375976971/tls.key\\\\\\\"\\\\nI1001 09:36:27.943129 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 09:36:27.955126 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 09:36:27.955153 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 09:36:27.955179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 09:36:27.955185 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 09:36:27.963955 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1001 09:36:27.963979 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1001 09:36:27.963986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964008 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 09:36:27.964015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 09:36:27.964018 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 09:36:27.964022 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1001 09:36:27.966112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.076007 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7aa5c938897ea621fc7c71c86e6a1dd2acca273e13afe71e9a27fc81f34cf2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.089561 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72701c6061105b7bb81545b5f0b035b220b22b2f8009e82809a037870c625685\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.096671 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.096712 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.096726 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.096747 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.096763 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:48Z","lastTransitionTime":"2025-10-01T09:36:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.101958 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.113531 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.125558 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc37698-af76-4818-b216-ffe35af67929\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1bb16edb25b63e957cc0bceb04a9fa9244e68881cb45be7c7d627c5faa74232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://995030421e0cbfeb7baa8ccb574d269dd178f8185bcaa9000fe5e7fc221b1188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d4c78dd607b81a28bcfe0ffa69c02628cf06aff5fd6a574069639c5bac044a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://144123d273f3f1af4fb857d731a16d9891b097c98229a035eb132f360fa8b2f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.136669 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"049825a6-0cec-4adf-8b77-4889bdc4d56f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18044055eb1f6cfd7cb05f330314231b1c563c34a51ba5ed79073988eb563cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc8192c84986df76c9a45deba61aaaa1b0a019fce0000e1c9916c4fe31d1ba1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-84q9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.147794 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.158786 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebc8a6e2b9c6104b0f773e99d68a62e1b2bf0adc05e6f291cf91e7a8cf1bc121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.199293 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.199561 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.199634 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.199703 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.199818 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:48Z","lastTransitionTime":"2025-10-01T09:36:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.301945 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.301979 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.301986 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.301998 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.302007 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:48Z","lastTransitionTime":"2025-10-01T09:36:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.405374 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.405438 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.405455 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.405481 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.405500 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:48Z","lastTransitionTime":"2025-10-01T09:36:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.509247 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.509287 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.509295 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.509311 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.509321 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:48Z","lastTransitionTime":"2025-10-01T09:36:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.522952 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.523206 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:36:48 crc kubenswrapper[4787]: E1001 09:36:48.523257 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.523530 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:36:48 crc kubenswrapper[4787]: E1001 09:36:48.523648 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:36:48 crc kubenswrapper[4787]: E1001 09:36:48.523899 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.571770 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.589448 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebc8a6e2b9c6104b0f773e99d68a62e1b2bf0adc05e6f291cf91e7a8cf1bc121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.607903 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"049825a6-0cec-4adf-8b77-4889bdc4d56f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18044055eb1f6cfd7cb05f330314231b1c563c34a51ba5ed79073988eb563cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc8192c84986df76c9a45deba61aaaa1b0a019fce0000e1c9916c4fe31d1ba1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-84q9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.617561 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.617630 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.617645 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.617682 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.617701 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:48Z","lastTransitionTime":"2025-10-01T09:36:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.630305 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4990da342a20f6ab1e97774294bd6856c49ca17de21d893a638c13c8e727dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.645475 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b093e96a244fdec4c5cdce62650d4da31c2ec46207a9c08bbd0a28bf286abb52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4664cb1dde25ada88989e81f0319aedc6dbd363f608a50623efaa3eeb8d00763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.657739 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d95f8b1aac879d1746d8a8af7e1d5e672b36c31d5b95c23fee70b237bc3716e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.669865 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.691402 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://899d7fc36a9f7a7335856902f4e1c915230f9b7e150d4139aeaab0652d74074c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899d7fc36a9f7a7335856902f4e1c915230f9b7e150d4139aeaab0652d74074c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:36:39Z\\\",\\\"message\\\":\\\"r/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1001 09:36:39.835210 6191 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nF1001 09:36:39.835211 6191 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:39Z is after 2025-08-24T17:21:41Z]\\\\nI1001 09:36:39.835216 6191 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1001 09:36:39.835221 6191 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1001 09:36:39.835225 6191 ovn.go:134] Ensuring zon\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dmsxm_openshift-ovn-kubernetes(eadf2545-886e-4642-b909-704c6cd4134e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.704438 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6zz6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb4a902-9efa-4088-a236-9a634fd085fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k5bxp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k5bxp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6zz6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.722201 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.722234 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.722245 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.722263 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.722274 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:48Z","lastTransitionTime":"2025-10-01T09:36:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.724360 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0814bed7-1117-4dff-97ef-a3cd1240e615\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c91542e7254c8289b734b3ffa8e8226f2445c915c264e09c179b99604abad1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:36:27Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1001 09:36:22.478145 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 09:36:22.478910 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3375976971/tls.crt::/tmp/serving-cert-3375976971/tls.key\\\\\\\"\\\\nI1001 09:36:27.943129 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 09:36:27.955126 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 09:36:27.955153 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 09:36:27.955179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 09:36:27.955185 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 09:36:27.963955 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1001 09:36:27.963979 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1001 09:36:27.963986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964008 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 09:36:27.964015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 09:36:27.964018 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 09:36:27.964022 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1001 09:36:27.966112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.737225 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7aa5c938897ea621fc7c71c86e6a1dd2acca273e13afe71e9a27fc81f34cf2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.757013 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c68b374f72c2a689eb09480e7ae3e88691daf0cb380a52f655eb1c651a66a281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.776967 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.792665 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.808638 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc37698-af76-4818-b216-ffe35af67929\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1bb16edb25b63e957cc0bceb04a9fa9244e68881cb45be7c7d627c5faa74232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://995030421e0cbfeb7baa8ccb574d269dd178f8185bcaa9000fe5e7fc221b1188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d4c78dd607b81a28bcfe0ffa69c02628cf06aff5fd6a574069639c5bac044a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://144123d273f3f1af4fb857d731a16d9891b097c98229a035eb132f360fa8b2f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.825183 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.825265 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.825291 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.825323 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.825347 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:48Z","lastTransitionTime":"2025-10-01T09:36:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.826901 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72701c6061105b7bb81545b5f0b035b220b22b2f8009e82809a037870c625685\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.928703 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.928770 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.928787 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.928815 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:48 crc kubenswrapper[4787]: I1001 09:36:48.928835 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:48Z","lastTransitionTime":"2025-10-01T09:36:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.032264 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.032322 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.032340 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.032366 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.032390 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:49Z","lastTransitionTime":"2025-10-01T09:36:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.135047 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.135201 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.135232 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.135268 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.135294 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:49Z","lastTransitionTime":"2025-10-01T09:36:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.240355 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.240761 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.240864 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.240963 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.241314 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:49Z","lastTransitionTime":"2025-10-01T09:36:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.345947 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.345993 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.346003 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.346020 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.346031 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:49Z","lastTransitionTime":"2025-10-01T09:36:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.448655 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.448696 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.448706 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.448720 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.448729 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:49Z","lastTransitionTime":"2025-10-01T09:36:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.523641 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:36:49 crc kubenswrapper[4787]: E1001 09:36:49.523767 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.551004 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.551040 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.551060 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.551091 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.551099 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:49Z","lastTransitionTime":"2025-10-01T09:36:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.653163 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.653198 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.653206 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.653220 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.653228 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:49Z","lastTransitionTime":"2025-10-01T09:36:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.755593 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.755639 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.755648 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.755669 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.755677 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:49Z","lastTransitionTime":"2025-10-01T09:36:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.858205 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.858243 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.858254 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.858268 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.858279 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:49Z","lastTransitionTime":"2025-10-01T09:36:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.960219 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.960269 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.960286 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.960307 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:49 crc kubenswrapper[4787]: I1001 09:36:49.960323 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:49Z","lastTransitionTime":"2025-10-01T09:36:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.062741 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.063018 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.063104 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.063212 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.063288 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:50Z","lastTransitionTime":"2025-10-01T09:36:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.166471 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.166530 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.166548 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.166571 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.166589 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:50Z","lastTransitionTime":"2025-10-01T09:36:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.168605 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0eb4a902-9efa-4088-a236-9a634fd085fd-metrics-certs\") pod \"network-metrics-daemon-6zz6g\" (UID: \"0eb4a902-9efa-4088-a236-9a634fd085fd\") " pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:36:50 crc kubenswrapper[4787]: E1001 09:36:50.169021 4787 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 09:36:50 crc kubenswrapper[4787]: E1001 09:36:50.169290 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0eb4a902-9efa-4088-a236-9a634fd085fd-metrics-certs podName:0eb4a902-9efa-4088-a236-9a634fd085fd nodeName:}" failed. No retries permitted until 2025-10-01 09:36:58.169258586 +0000 UTC m=+50.284402773 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0eb4a902-9efa-4088-a236-9a634fd085fd-metrics-certs") pod "network-metrics-daemon-6zz6g" (UID: "0eb4a902-9efa-4088-a236-9a634fd085fd") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.269826 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.269909 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.269936 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.269969 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.270007 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:50Z","lastTransitionTime":"2025-10-01T09:36:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.372356 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.372409 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.372421 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.372439 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.372451 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:50Z","lastTransitionTime":"2025-10-01T09:36:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.474388 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.474434 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.474446 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.474463 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.474476 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:50Z","lastTransitionTime":"2025-10-01T09:36:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.523046 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:36:50 crc kubenswrapper[4787]: E1001 09:36:50.523181 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.523049 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.523046 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:36:50 crc kubenswrapper[4787]: E1001 09:36:50.523260 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:36:50 crc kubenswrapper[4787]: E1001 09:36:50.523466 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.576453 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.576490 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.576501 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.576519 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.576537 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:50Z","lastTransitionTime":"2025-10-01T09:36:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.678846 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.679120 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.679205 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.679318 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.679421 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:50Z","lastTransitionTime":"2025-10-01T09:36:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.782658 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.782705 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.782721 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.782741 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.782754 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:50Z","lastTransitionTime":"2025-10-01T09:36:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.885676 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.885941 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.886020 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.886125 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.886194 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:50Z","lastTransitionTime":"2025-10-01T09:36:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.989055 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.989378 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.989442 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.989503 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:50 crc kubenswrapper[4787]: I1001 09:36:50.989560 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:50Z","lastTransitionTime":"2025-10-01T09:36:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.092212 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.092468 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.092573 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.092661 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.092746 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:51Z","lastTransitionTime":"2025-10-01T09:36:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.195576 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.195620 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.195633 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.195652 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.195663 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:51Z","lastTransitionTime":"2025-10-01T09:36:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.298353 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.298397 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.298409 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.298424 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.298436 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:51Z","lastTransitionTime":"2025-10-01T09:36:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.401023 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.401112 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.401125 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.401145 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.401158 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:51Z","lastTransitionTime":"2025-10-01T09:36:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.503290 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.503321 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.503343 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.503357 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.503368 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:51Z","lastTransitionTime":"2025-10-01T09:36:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.523312 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:36:51 crc kubenswrapper[4787]: E1001 09:36:51.523513 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.605281 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.605324 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.605334 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.605349 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.605359 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:51Z","lastTransitionTime":"2025-10-01T09:36:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.707526 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.707553 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.707561 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.707573 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.707581 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:51Z","lastTransitionTime":"2025-10-01T09:36:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.809379 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.809412 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.809420 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.809432 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.809441 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:51Z","lastTransitionTime":"2025-10-01T09:36:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.911649 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.911866 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.911939 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.912006 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:51 crc kubenswrapper[4787]: I1001 09:36:51.912066 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:51Z","lastTransitionTime":"2025-10-01T09:36:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.013791 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.014029 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.014131 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.014210 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.014284 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:52Z","lastTransitionTime":"2025-10-01T09:36:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.116425 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.116484 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.116494 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.116506 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.116515 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:52Z","lastTransitionTime":"2025-10-01T09:36:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.218565 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.218883 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.219160 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.219396 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.219490 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:52Z","lastTransitionTime":"2025-10-01T09:36:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.321799 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.321830 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.321837 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.321851 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.321862 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:52Z","lastTransitionTime":"2025-10-01T09:36:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.424031 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.424110 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.424144 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.424223 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.424237 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:52Z","lastTransitionTime":"2025-10-01T09:36:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.523445 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:36:52 crc kubenswrapper[4787]: E1001 09:36:52.523571 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.523587 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:36:52 crc kubenswrapper[4787]: E1001 09:36:52.523682 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.523465 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:36:52 crc kubenswrapper[4787]: E1001 09:36:52.523776 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.526992 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.527059 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.527090 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.527107 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.527117 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:52Z","lastTransitionTime":"2025-10-01T09:36:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.629009 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.629044 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.629057 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.629070 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.629118 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:52Z","lastTransitionTime":"2025-10-01T09:36:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.731692 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.731928 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.732005 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.732093 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.732154 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:52Z","lastTransitionTime":"2025-10-01T09:36:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.834662 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.834746 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.834764 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.834788 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.834806 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:52Z","lastTransitionTime":"2025-10-01T09:36:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.937789 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.937820 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.937830 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.937847 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:52 crc kubenswrapper[4787]: I1001 09:36:52.937858 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:52Z","lastTransitionTime":"2025-10-01T09:36:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.040229 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.040457 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.040551 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.040620 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.040696 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:53Z","lastTransitionTime":"2025-10-01T09:36:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.143538 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.143955 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.144022 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.144120 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.144255 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:53Z","lastTransitionTime":"2025-10-01T09:36:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.247439 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.247479 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.247491 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.247507 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.247533 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:53Z","lastTransitionTime":"2025-10-01T09:36:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.350494 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.350541 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.350553 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.350572 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.350585 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:53Z","lastTransitionTime":"2025-10-01T09:36:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.453118 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.453163 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.453172 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.453187 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.453197 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:53Z","lastTransitionTime":"2025-10-01T09:36:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.523735 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:36:53 crc kubenswrapper[4787]: E1001 09:36:53.523861 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.555200 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.555268 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.555284 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.555301 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.555313 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:53Z","lastTransitionTime":"2025-10-01T09:36:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.657950 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.657988 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.658001 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.658016 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.658028 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:53Z","lastTransitionTime":"2025-10-01T09:36:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.760949 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.760995 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.761005 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.761023 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.761038 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:53Z","lastTransitionTime":"2025-10-01T09:36:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.863627 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.864157 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.864234 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.864328 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.864424 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:53Z","lastTransitionTime":"2025-10-01T09:36:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.967415 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.967628 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.967722 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.967788 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:53 crc kubenswrapper[4787]: I1001 09:36:53.967855 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:53Z","lastTransitionTime":"2025-10-01T09:36:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.069685 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.069748 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.069760 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.069775 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.069786 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:54Z","lastTransitionTime":"2025-10-01T09:36:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.172405 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.172531 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.172550 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.172570 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.172587 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:54Z","lastTransitionTime":"2025-10-01T09:36:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.274878 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.274914 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.274922 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.274943 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.274953 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:54Z","lastTransitionTime":"2025-10-01T09:36:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.377619 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.377667 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.377677 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.377695 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.377708 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:54Z","lastTransitionTime":"2025-10-01T09:36:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.479697 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.479729 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.479739 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.479754 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.479765 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:54Z","lastTransitionTime":"2025-10-01T09:36:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.523633 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.523716 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:36:54 crc kubenswrapper[4787]: E1001 09:36:54.523765 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.523815 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:36:54 crc kubenswrapper[4787]: E1001 09:36:54.523883 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:36:54 crc kubenswrapper[4787]: E1001 09:36:54.523943 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.581464 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.581497 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.581506 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.581519 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.581528 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:54Z","lastTransitionTime":"2025-10-01T09:36:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.683348 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.683393 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.683405 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.683426 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.683442 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:54Z","lastTransitionTime":"2025-10-01T09:36:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.730264 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.730299 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.730307 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.730320 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.730330 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:54Z","lastTransitionTime":"2025-10-01T09:36:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:54 crc kubenswrapper[4787]: E1001 09:36:54.744192 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:54Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.749014 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.749063 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.749089 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.749104 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.749113 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:54Z","lastTransitionTime":"2025-10-01T09:36:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:54 crc kubenswrapper[4787]: E1001 09:36:54.763680 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:54Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.767920 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.767958 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.767969 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.767985 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.767994 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:54Z","lastTransitionTime":"2025-10-01T09:36:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:54 crc kubenswrapper[4787]: E1001 09:36:54.780559 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:54Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.784500 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.784534 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.784543 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.784556 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.784566 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:54Z","lastTransitionTime":"2025-10-01T09:36:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:54 crc kubenswrapper[4787]: E1001 09:36:54.795355 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:54Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.799168 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.799247 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.799260 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.799277 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.799317 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:54Z","lastTransitionTime":"2025-10-01T09:36:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:54 crc kubenswrapper[4787]: E1001 09:36:54.813250 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:36:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:54Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:54 crc kubenswrapper[4787]: E1001 09:36:54.813415 4787 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.815524 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.815631 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.815701 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.815782 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.815877 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:54Z","lastTransitionTime":"2025-10-01T09:36:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.918426 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.918466 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.918480 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.918498 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:54 crc kubenswrapper[4787]: I1001 09:36:54.918513 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:54Z","lastTransitionTime":"2025-10-01T09:36:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.021023 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.021178 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.021194 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.021213 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.021227 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:55Z","lastTransitionTime":"2025-10-01T09:36:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.123200 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.123520 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.123676 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.123784 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.123872 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:55Z","lastTransitionTime":"2025-10-01T09:36:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.226373 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.226406 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.226416 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.226432 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.226444 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:55Z","lastTransitionTime":"2025-10-01T09:36:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.328780 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.328814 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.328822 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.328836 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.328846 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:55Z","lastTransitionTime":"2025-10-01T09:36:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.431587 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.431639 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.431659 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.431684 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.431700 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:55Z","lastTransitionTime":"2025-10-01T09:36:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.523929 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:36:55 crc kubenswrapper[4787]: E1001 09:36:55.524231 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.525380 4787 scope.go:117] "RemoveContainer" containerID="899d7fc36a9f7a7335856902f4e1c915230f9b7e150d4139aeaab0652d74074c" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.534949 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.535211 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.535324 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.535438 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.535541 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:55Z","lastTransitionTime":"2025-10-01T09:36:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.643227 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.643786 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.643803 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.643828 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.643842 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:55Z","lastTransitionTime":"2025-10-01T09:36:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.746656 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.746745 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.746784 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.746809 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.746827 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:55Z","lastTransitionTime":"2025-10-01T09:36:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.829570 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dmsxm_eadf2545-886e-4642-b909-704c6cd4134e/ovnkube-controller/1.log" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.833470 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" event={"ID":"eadf2545-886e-4642-b909-704c6cd4134e","Type":"ContainerStarted","Data":"e77339fe976956bb53dcd610364af111550cb6bfa1103daf7f706a9ee74c4439"} Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.834000 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.850320 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.850387 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.850404 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.850430 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.850447 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:55Z","lastTransitionTime":"2025-10-01T09:36:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.854841 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.869718 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebc8a6e2b9c6104b0f773e99d68a62e1b2bf0adc05e6f291cf91e7a8cf1bc121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.887158 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"049825a6-0cec-4adf-8b77-4889bdc4d56f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18044055eb1f6cfd7cb05f330314231b1c563c34a51ba5ed79073988eb563cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc8192c84986df76c9a45deba61aaaa1b0a019fce0000e1c9916c4fe31d1ba1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-84q9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.905934 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.931416 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4990da342a20f6ab1e97774294bd6856c49ca17de21d893a638c13c8e727dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.945265 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b093e96a244fdec4c5cdce62650d4da31c2ec46207a9c08bbd0a28bf286abb52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4664cb1dde25ada88989e81f0319aedc6dbd363f608a50623efaa3eeb8d00763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.955413 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.955477 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.955490 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.955517 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.955538 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:55Z","lastTransitionTime":"2025-10-01T09:36:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.960847 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d95f8b1aac879d1746d8a8af7e1d5e672b36c31d5b95c23fee70b237bc3716e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.978099 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0814bed7-1117-4dff-97ef-a3cd1240e615\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c91542e7254c8289b734b3ffa8e8226f2445c915c264e09c179b99604abad1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:36:27Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1001 09:36:22.478145 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 09:36:22.478910 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3375976971/tls.crt::/tmp/serving-cert-3375976971/tls.key\\\\\\\"\\\\nI1001 09:36:27.943129 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 09:36:27.955126 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 09:36:27.955153 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 09:36:27.955179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 09:36:27.955185 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 09:36:27.963955 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1001 09:36:27.963979 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1001 09:36:27.963986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964008 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 09:36:27.964015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 09:36:27.964018 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 09:36:27.964022 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1001 09:36:27.966112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:55 crc kubenswrapper[4787]: I1001 09:36:55.988803 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7aa5c938897ea621fc7c71c86e6a1dd2acca273e13afe71e9a27fc81f34cf2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:55Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.004394 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c68b374f72c2a689eb09480e7ae3e88691daf0cb380a52f655eb1c651a66a281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.035599 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e77339fe976956bb53dcd610364af111550cb6bfa1103daf7f706a9ee74c4439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899d7fc36a9f7a7335856902f4e1c915230f9b7e150d4139aeaab0652d74074c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:36:39Z\\\",\\\"message\\\":\\\"r/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1001 09:36:39.835210 6191 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nF1001 09:36:39.835211 6191 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:39Z is after 2025-08-24T17:21:41Z]\\\\nI1001 09:36:39.835216 6191 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1001 09:36:39.835221 6191 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1001 09:36:39.835225 6191 ovn.go:134] Ensuring zon\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.051737 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6zz6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb4a902-9efa-4088-a236-9a634fd085fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k5bxp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k5bxp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6zz6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.058203 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.058405 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.058484 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.058564 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.058646 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:56Z","lastTransitionTime":"2025-10-01T09:36:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.069325 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc37698-af76-4818-b216-ffe35af67929\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1bb16edb25b63e957cc0bceb04a9fa9244e68881cb45be7c7d627c5faa74232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://995030421e0cbfeb7baa8ccb574d269dd178f8185bcaa9000fe5e7fc221b1188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d4c78dd607b81a28bcfe0ffa69c02628cf06aff5fd6a574069639c5bac044a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://144123d273f3f1af4fb857d731a16d9891b097c98229a035eb132f360fa8b2f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.094310 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72701c6061105b7bb81545b5f0b035b220b22b2f8009e82809a037870c625685\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.112295 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.129166 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.163136 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.163475 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.163553 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.163701 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.163793 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:56Z","lastTransitionTime":"2025-10-01T09:36:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.266859 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.266917 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.266931 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.266954 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.266969 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:56Z","lastTransitionTime":"2025-10-01T09:36:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.369575 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.369620 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.369631 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.369648 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.369659 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:56Z","lastTransitionTime":"2025-10-01T09:36:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.473755 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.474183 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.474288 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.474435 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.474536 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:56Z","lastTransitionTime":"2025-10-01T09:36:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.523194 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.523284 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:36:56 crc kubenswrapper[4787]: E1001 09:36:56.523337 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.523283 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:36:56 crc kubenswrapper[4787]: E1001 09:36:56.523661 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:36:56 crc kubenswrapper[4787]: E1001 09:36:56.523534 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.578504 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.578556 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.578569 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.578592 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.578603 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:56Z","lastTransitionTime":"2025-10-01T09:36:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.681470 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.681530 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.681545 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.681569 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.681584 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:56Z","lastTransitionTime":"2025-10-01T09:36:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.784005 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.784067 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.784109 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.784137 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.784152 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:56Z","lastTransitionTime":"2025-10-01T09:36:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.838958 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dmsxm_eadf2545-886e-4642-b909-704c6cd4134e/ovnkube-controller/2.log" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.839717 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dmsxm_eadf2545-886e-4642-b909-704c6cd4134e/ovnkube-controller/1.log" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.843067 4787 generic.go:334] "Generic (PLEG): container finished" podID="eadf2545-886e-4642-b909-704c6cd4134e" containerID="e77339fe976956bb53dcd610364af111550cb6bfa1103daf7f706a9ee74c4439" exitCode=1 Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.843130 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" event={"ID":"eadf2545-886e-4642-b909-704c6cd4134e","Type":"ContainerDied","Data":"e77339fe976956bb53dcd610364af111550cb6bfa1103daf7f706a9ee74c4439"} Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.843178 4787 scope.go:117] "RemoveContainer" containerID="899d7fc36a9f7a7335856902f4e1c915230f9b7e150d4139aeaab0652d74074c" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.844438 4787 scope.go:117] "RemoveContainer" containerID="e77339fe976956bb53dcd610364af111550cb6bfa1103daf7f706a9ee74c4439" Oct 01 09:36:56 crc kubenswrapper[4787]: E1001 09:36:56.844832 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dmsxm_openshift-ovn-kubernetes(eadf2545-886e-4642-b909-704c6cd4134e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" podUID="eadf2545-886e-4642-b909-704c6cd4134e" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.870940 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0814bed7-1117-4dff-97ef-a3cd1240e615\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c91542e7254c8289b734b3ffa8e8226f2445c915c264e09c179b99604abad1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:36:27Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1001 09:36:22.478145 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 09:36:22.478910 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3375976971/tls.crt::/tmp/serving-cert-3375976971/tls.key\\\\\\\"\\\\nI1001 09:36:27.943129 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 09:36:27.955126 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 09:36:27.955153 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 09:36:27.955179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 09:36:27.955185 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 09:36:27.963955 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1001 09:36:27.963979 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1001 09:36:27.963986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964008 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 09:36:27.964015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 09:36:27.964018 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 09:36:27.964022 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1001 09:36:27.966112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.886939 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.887250 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.887352 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.887456 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.887542 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:56Z","lastTransitionTime":"2025-10-01T09:36:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.890050 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7aa5c938897ea621fc7c71c86e6a1dd2acca273e13afe71e9a27fc81f34cf2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.916184 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c68b374f72c2a689eb09480e7ae3e88691daf0cb380a52f655eb1c651a66a281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.950191 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e77339fe976956bb53dcd610364af111550cb6bfa1103daf7f706a9ee74c4439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899d7fc36a9f7a7335856902f4e1c915230f9b7e150d4139aeaab0652d74074c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:36:39Z\\\",\\\"message\\\":\\\"r/kube-apiserver-crc after 0 failed attempt(s)\\\\nI1001 09:36:39.835210 6191 default_network_controller.go:776] Recording success event on pod openshift-kube-apiserver/kube-apiserver-crc\\\\nF1001 09:36:39.835211 6191 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:39Z is after 2025-08-24T17:21:41Z]\\\\nI1001 09:36:39.835216 6191 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1001 09:36:39.835221 6191 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1001 09:36:39.835225 6191 ovn.go:134] Ensuring zon\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e77339fe976956bb53dcd610364af111550cb6bfa1103daf7f706a9ee74c4439\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:36:56Z\\\",\\\"message\\\":\\\"{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: kube-storage-version-migrator-operator,},ClusterIP:10.217.5.36,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.36],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1001 09:36:56.575330 6417 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.965608 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6zz6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb4a902-9efa-4088-a236-9a634fd085fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k5bxp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k5bxp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6zz6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.981060 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc37698-af76-4818-b216-ffe35af67929\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1bb16edb25b63e957cc0bceb04a9fa9244e68881cb45be7c7d627c5faa74232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://995030421e0cbfeb7baa8ccb574d269dd178f8185bcaa9000fe5e7fc221b1188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d4c78dd607b81a28bcfe0ffa69c02628cf06aff5fd6a574069639c5bac044a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://144123d273f3f1af4fb857d731a16d9891b097c98229a035eb132f360fa8b2f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.989731 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.989797 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.989810 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.989833 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.989849 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:56Z","lastTransitionTime":"2025-10-01T09:36:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:56 crc kubenswrapper[4787]: I1001 09:36:56.995950 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72701c6061105b7bb81545b5f0b035b220b22b2f8009e82809a037870c625685\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.013847 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.029491 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.049194 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.066591 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebc8a6e2b9c6104b0f773e99d68a62e1b2bf0adc05e6f291cf91e7a8cf1bc121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.080274 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"049825a6-0cec-4adf-8b77-4889bdc4d56f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18044055eb1f6cfd7cb05f330314231b1c563c34a51ba5ed79073988eb563cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc8192c84986df76c9a45deba61aaaa1b0a019fce0000e1c9916c4fe31d1ba1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-84q9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.092717 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.092768 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.092780 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.092799 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.092813 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:57Z","lastTransitionTime":"2025-10-01T09:36:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.102279 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.120987 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4990da342a20f6ab1e97774294bd6856c49ca17de21d893a638c13c8e727dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.141524 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b093e96a244fdec4c5cdce62650d4da31c2ec46207a9c08bbd0a28bf286abb52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4664cb1dde25ada88989e81f0319aedc6dbd363f608a50623efaa3eeb8d00763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.158677 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d95f8b1aac879d1746d8a8af7e1d5e672b36c31d5b95c23fee70b237bc3716e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.195550 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.195607 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.195619 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.195636 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.195648 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:57Z","lastTransitionTime":"2025-10-01T09:36:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.297282 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.297340 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.297357 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.297381 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.297399 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:57Z","lastTransitionTime":"2025-10-01T09:36:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.399661 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.399704 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.399715 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.399729 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.399739 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:57Z","lastTransitionTime":"2025-10-01T09:36:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.501983 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.502022 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.502034 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.502050 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.502059 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:57Z","lastTransitionTime":"2025-10-01T09:36:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.522783 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:36:57 crc kubenswrapper[4787]: E1001 09:36:57.522964 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.604059 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.604184 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.604204 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.604230 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.604247 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:57Z","lastTransitionTime":"2025-10-01T09:36:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.706735 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.706783 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.706794 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.706809 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.706820 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:57Z","lastTransitionTime":"2025-10-01T09:36:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.810023 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.810091 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.810139 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.810160 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.810170 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:57Z","lastTransitionTime":"2025-10-01T09:36:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.848687 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dmsxm_eadf2545-886e-4642-b909-704c6cd4134e/ovnkube-controller/2.log" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.854859 4787 scope.go:117] "RemoveContainer" containerID="e77339fe976956bb53dcd610364af111550cb6bfa1103daf7f706a9ee74c4439" Oct 01 09:36:57 crc kubenswrapper[4787]: E1001 09:36:57.855236 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dmsxm_openshift-ovn-kubernetes(eadf2545-886e-4642-b909-704c6cd4134e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" podUID="eadf2545-886e-4642-b909-704c6cd4134e" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.870368 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0814bed7-1117-4dff-97ef-a3cd1240e615\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c91542e7254c8289b734b3ffa8e8226f2445c915c264e09c179b99604abad1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:36:27Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1001 09:36:22.478145 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 09:36:22.478910 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3375976971/tls.crt::/tmp/serving-cert-3375976971/tls.key\\\\\\\"\\\\nI1001 09:36:27.943129 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 09:36:27.955126 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 09:36:27.955153 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 09:36:27.955179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 09:36:27.955185 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 09:36:27.963955 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1001 09:36:27.963979 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1001 09:36:27.963986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964008 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 09:36:27.964015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 09:36:27.964018 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 09:36:27.964022 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1001 09:36:27.966112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.881776 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7aa5c938897ea621fc7c71c86e6a1dd2acca273e13afe71e9a27fc81f34cf2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.895655 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c68b374f72c2a689eb09480e7ae3e88691daf0cb380a52f655eb1c651a66a281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.913128 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.913191 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.913201 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.913218 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.913253 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:57Z","lastTransitionTime":"2025-10-01T09:36:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.917306 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e77339fe976956bb53dcd610364af111550cb6bfa1103daf7f706a9ee74c4439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e77339fe976956bb53dcd610364af111550cb6bfa1103daf7f706a9ee74c4439\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:36:56Z\\\",\\\"message\\\":\\\"{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: kube-storage-version-migrator-operator,},ClusterIP:10.217.5.36,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.36],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1001 09:36:56.575330 6417 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dmsxm_openshift-ovn-kubernetes(eadf2545-886e-4642-b909-704c6cd4134e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.931998 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6zz6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb4a902-9efa-4088-a236-9a634fd085fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k5bxp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k5bxp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6zz6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.948586 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc37698-af76-4818-b216-ffe35af67929\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1bb16edb25b63e957cc0bceb04a9fa9244e68881cb45be7c7d627c5faa74232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://995030421e0cbfeb7baa8ccb574d269dd178f8185bcaa9000fe5e7fc221b1188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d4c78dd607b81a28bcfe0ffa69c02628cf06aff5fd6a574069639c5bac044a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://144123d273f3f1af4fb857d731a16d9891b097c98229a035eb132f360fa8b2f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.965110 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72701c6061105b7bb81545b5f0b035b220b22b2f8009e82809a037870c625685\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.982503 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:57 crc kubenswrapper[4787]: I1001 09:36:57.997532 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:57Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.016265 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:58Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.016611 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.016652 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.016664 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.016682 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.016695 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:58Z","lastTransitionTime":"2025-10-01T09:36:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.029969 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebc8a6e2b9c6104b0f773e99d68a62e1b2bf0adc05e6f291cf91e7a8cf1bc121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:58Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.043782 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"049825a6-0cec-4adf-8b77-4889bdc4d56f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18044055eb1f6cfd7cb05f330314231b1c563c34a51ba5ed79073988eb563cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc8192c84986df76c9a45deba61aaaa1b0a019fce0000e1c9916c4fe31d1ba1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-84q9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:58Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.056371 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:58Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.074953 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4990da342a20f6ab1e97774294bd6856c49ca17de21d893a638c13c8e727dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:58Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.091007 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b093e96a244fdec4c5cdce62650d4da31c2ec46207a9c08bbd0a28bf286abb52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4664cb1dde25ada88989e81f0319aedc6dbd363f608a50623efaa3eeb8d00763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:58Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.105106 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d95f8b1aac879d1746d8a8af7e1d5e672b36c31d5b95c23fee70b237bc3716e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:58Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.119257 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.119328 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.119345 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.119376 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.119392 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:58Z","lastTransitionTime":"2025-10-01T09:36:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.222679 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.222740 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.222754 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.222778 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.222792 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:58Z","lastTransitionTime":"2025-10-01T09:36:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.256750 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0eb4a902-9efa-4088-a236-9a634fd085fd-metrics-certs\") pod \"network-metrics-daemon-6zz6g\" (UID: \"0eb4a902-9efa-4088-a236-9a634fd085fd\") " pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:36:58 crc kubenswrapper[4787]: E1001 09:36:58.256937 4787 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 09:36:58 crc kubenswrapper[4787]: E1001 09:36:58.257013 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0eb4a902-9efa-4088-a236-9a634fd085fd-metrics-certs podName:0eb4a902-9efa-4088-a236-9a634fd085fd nodeName:}" failed. No retries permitted until 2025-10-01 09:37:14.25699159 +0000 UTC m=+66.372135747 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0eb4a902-9efa-4088-a236-9a634fd085fd-metrics-certs") pod "network-metrics-daemon-6zz6g" (UID: "0eb4a902-9efa-4088-a236-9a634fd085fd") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.325802 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.325842 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.325852 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.325867 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.325878 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:58Z","lastTransitionTime":"2025-10-01T09:36:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.428972 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.429019 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.429032 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.429053 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.429065 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:58Z","lastTransitionTime":"2025-10-01T09:36:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.523108 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.523139 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.523199 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:36:58 crc kubenswrapper[4787]: E1001 09:36:58.523319 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:36:58 crc kubenswrapper[4787]: E1001 09:36:58.523399 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:36:58 crc kubenswrapper[4787]: E1001 09:36:58.523555 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.532346 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.532388 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.532396 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.532414 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.532424 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:58Z","lastTransitionTime":"2025-10-01T09:36:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.535554 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b093e96a244fdec4c5cdce62650d4da31c2ec46207a9c08bbd0a28bf286abb52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4664cb1dde25ada88989e81f0319aedc6dbd363f608a50623efaa3eeb8d00763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:58Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.545890 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d95f8b1aac879d1746d8a8af7e1d5e672b36c31d5b95c23fee70b237bc3716e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:58Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.555824 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:58Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.566068 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4990da342a20f6ab1e97774294bd6856c49ca17de21d893a638c13c8e727dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:58Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.575150 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6zz6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb4a902-9efa-4088-a236-9a634fd085fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k5bxp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k5bxp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6zz6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:58Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.586319 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0814bed7-1117-4dff-97ef-a3cd1240e615\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c91542e7254c8289b734b3ffa8e8226f2445c915c264e09c179b99604abad1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:36:27Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1001 09:36:22.478145 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 09:36:22.478910 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3375976971/tls.crt::/tmp/serving-cert-3375976971/tls.key\\\\\\\"\\\\nI1001 09:36:27.943129 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 09:36:27.955126 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 09:36:27.955153 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 09:36:27.955179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 09:36:27.955185 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 09:36:27.963955 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1001 09:36:27.963979 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1001 09:36:27.963986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964008 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 09:36:27.964015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 09:36:27.964018 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 09:36:27.964022 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1001 09:36:27.966112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:58Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.595658 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7aa5c938897ea621fc7c71c86e6a1dd2acca273e13afe71e9a27fc81f34cf2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:58Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.607278 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c68b374f72c2a689eb09480e7ae3e88691daf0cb380a52f655eb1c651a66a281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:58Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.625863 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e77339fe976956bb53dcd610364af111550cb6bfa1103daf7f706a9ee74c4439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e77339fe976956bb53dcd610364af111550cb6bfa1103daf7f706a9ee74c4439\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:36:56Z\\\",\\\"message\\\":\\\"{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: kube-storage-version-migrator-operator,},ClusterIP:10.217.5.36,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.36],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1001 09:36:56.575330 6417 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dmsxm_openshift-ovn-kubernetes(eadf2545-886e-4642-b909-704c6cd4134e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:58Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.634137 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.634185 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.634200 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.634229 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.634244 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:58Z","lastTransitionTime":"2025-10-01T09:36:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.641642 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:58Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.653784 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc37698-af76-4818-b216-ffe35af67929\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1bb16edb25b63e957cc0bceb04a9fa9244e68881cb45be7c7d627c5faa74232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://995030421e0cbfeb7baa8ccb574d269dd178f8185bcaa9000fe5e7fc221b1188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d4c78dd607b81a28bcfe0ffa69c02628cf06aff5fd6a574069639c5bac044a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://144123d273f3f1af4fb857d731a16d9891b097c98229a035eb132f360fa8b2f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:58Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.664557 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72701c6061105b7bb81545b5f0b035b220b22b2f8009e82809a037870c625685\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:58Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.674405 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:58Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.684459 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:58Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.695196 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebc8a6e2b9c6104b0f773e99d68a62e1b2bf0adc05e6f291cf91e7a8cf1bc121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:58Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.705432 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"049825a6-0cec-4adf-8b77-4889bdc4d56f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18044055eb1f6cfd7cb05f330314231b1c563c34a51ba5ed79073988eb563cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc8192c84986df76c9a45deba61aaaa1b0a019fce0000e1c9916c4fe31d1ba1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-84q9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:58Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.736446 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.736507 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.736518 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.736535 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.736548 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:58Z","lastTransitionTime":"2025-10-01T09:36:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.838808 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.838847 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.838858 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.838873 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.838882 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:58Z","lastTransitionTime":"2025-10-01T09:36:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.942023 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.942061 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.942069 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.942113 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:58 crc kubenswrapper[4787]: I1001 09:36:58.942122 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:58Z","lastTransitionTime":"2025-10-01T09:36:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.045718 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.047834 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.047869 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.047903 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.047923 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:59Z","lastTransitionTime":"2025-10-01T09:36:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.150618 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.150680 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.150692 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.150708 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.150717 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:59Z","lastTransitionTime":"2025-10-01T09:36:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.253601 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.253649 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.253670 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.253688 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.253697 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:59Z","lastTransitionTime":"2025-10-01T09:36:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.356120 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.356157 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.356165 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.356178 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.356187 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:59Z","lastTransitionTime":"2025-10-01T09:36:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.458848 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.458896 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.458907 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.458922 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.458932 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:59Z","lastTransitionTime":"2025-10-01T09:36:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.523497 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:36:59 crc kubenswrapper[4787]: E1001 09:36:59.523676 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.562298 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.562360 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.562380 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.562398 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.562409 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:59Z","lastTransitionTime":"2025-10-01T09:36:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.666248 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.666276 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.666284 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.666297 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.666306 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:59Z","lastTransitionTime":"2025-10-01T09:36:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.768424 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.768464 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.768473 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.768487 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.768496 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:59Z","lastTransitionTime":"2025-10-01T09:36:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.843631 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.860392 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.863057 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:59Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.873255 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.873342 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.873360 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.873389 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.873407 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:59Z","lastTransitionTime":"2025-10-01T09:36:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.873933 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebc8a6e2b9c6104b0f773e99d68a62e1b2bf0adc05e6f291cf91e7a8cf1bc121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:59Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.888243 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"049825a6-0cec-4adf-8b77-4889bdc4d56f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18044055eb1f6cfd7cb05f330314231b1c563c34a51ba5ed79073988eb563cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc8192c84986df76c9a45deba61aaaa1b0a019fce0000e1c9916c4fe31d1ba1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-84q9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:59Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.902408 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4990da342a20f6ab1e97774294bd6856c49ca17de21d893a638c13c8e727dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:59Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.915350 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b093e96a244fdec4c5cdce62650d4da31c2ec46207a9c08bbd0a28bf286abb52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4664cb1dde25ada88989e81f0319aedc6dbd363f608a50623efaa3eeb8d00763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:59Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.926898 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d95f8b1aac879d1746d8a8af7e1d5e672b36c31d5b95c23fee70b237bc3716e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:59Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.939456 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:59Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.956913 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e77339fe976956bb53dcd610364af111550cb6bfa1103daf7f706a9ee74c4439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e77339fe976956bb53dcd610364af111550cb6bfa1103daf7f706a9ee74c4439\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:36:56Z\\\",\\\"message\\\":\\\"{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: kube-storage-version-migrator-operator,},ClusterIP:10.217.5.36,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.36],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1001 09:36:56.575330 6417 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dmsxm_openshift-ovn-kubernetes(eadf2545-886e-4642-b909-704c6cd4134e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:59Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.976825 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.976880 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.976891 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.976907 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.976918 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:36:59Z","lastTransitionTime":"2025-10-01T09:36:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.976909 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6zz6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb4a902-9efa-4088-a236-9a634fd085fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k5bxp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k5bxp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6zz6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:59Z is after 2025-08-24T17:21:41Z" Oct 01 09:36:59 crc kubenswrapper[4787]: I1001 09:36:59.993490 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0814bed7-1117-4dff-97ef-a3cd1240e615\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c91542e7254c8289b734b3ffa8e8226f2445c915c264e09c179b99604abad1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:36:27Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1001 09:36:22.478145 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 09:36:22.478910 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3375976971/tls.crt::/tmp/serving-cert-3375976971/tls.key\\\\\\\"\\\\nI1001 09:36:27.943129 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 09:36:27.955126 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 09:36:27.955153 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 09:36:27.955179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 09:36:27.955185 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 09:36:27.963955 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1001 09:36:27.963979 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1001 09:36:27.963986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964008 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 09:36:27.964015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 09:36:27.964018 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 09:36:27.964022 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1001 09:36:27.966112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:36:59Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.007527 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7aa5c938897ea621fc7c71c86e6a1dd2acca273e13afe71e9a27fc81f34cf2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:00Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.025706 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c68b374f72c2a689eb09480e7ae3e88691daf0cb380a52f655eb1c651a66a281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:00Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.040282 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:00Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.056341 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:00Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.076842 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc37698-af76-4818-b216-ffe35af67929\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1bb16edb25b63e957cc0bceb04a9fa9244e68881cb45be7c7d627c5faa74232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://995030421e0cbfeb7baa8ccb574d269dd178f8185bcaa9000fe5e7fc221b1188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d4c78dd607b81a28bcfe0ffa69c02628cf06aff5fd6a574069639c5bac044a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://144123d273f3f1af4fb857d731a16d9891b097c98229a035eb132f360fa8b2f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:00Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.079671 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.079981 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.080005 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.080038 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.080059 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:00Z","lastTransitionTime":"2025-10-01T09:37:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.095463 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72701c6061105b7bb81545b5f0b035b220b22b2f8009e82809a037870c625685\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:00Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.183894 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.183950 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.183960 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.183975 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.183987 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:00Z","lastTransitionTime":"2025-10-01T09:37:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.286481 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.286563 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.286592 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.286620 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.286637 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:00Z","lastTransitionTime":"2025-10-01T09:37:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.381225 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:37:00 crc kubenswrapper[4787]: E1001 09:37:00.381472 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:37:32.381435631 +0000 UTC m=+84.496579788 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.381594 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.381631 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.381671 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.381727 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:37:00 crc kubenswrapper[4787]: E1001 09:37:00.381825 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 09:37:00 crc kubenswrapper[4787]: E1001 09:37:00.381877 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 09:37:00 crc kubenswrapper[4787]: E1001 09:37:00.381886 4787 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 09:37:00 crc kubenswrapper[4787]: E1001 09:37:00.381907 4787 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:37:00 crc kubenswrapper[4787]: E1001 09:37:00.381935 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 09:37:32.381926173 +0000 UTC m=+84.497070330 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 09:37:00 crc kubenswrapper[4787]: E1001 09:37:00.381872 4787 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 09:37:00 crc kubenswrapper[4787]: E1001 09:37:00.381986 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 09:37:32.381977144 +0000 UTC m=+84.497121291 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 09:37:00 crc kubenswrapper[4787]: E1001 09:37:00.381994 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 09:37:00 crc kubenswrapper[4787]: E1001 09:37:00.382057 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 09:37:00 crc kubenswrapper[4787]: E1001 09:37:00.382101 4787 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:37:00 crc kubenswrapper[4787]: E1001 09:37:00.382003 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-01 09:37:32.381995464 +0000 UTC m=+84.497139621 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:37:00 crc kubenswrapper[4787]: E1001 09:37:00.382191 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-01 09:37:32.382166328 +0000 UTC m=+84.497310545 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.389536 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.389571 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.389581 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.389594 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.389605 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:00Z","lastTransitionTime":"2025-10-01T09:37:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.492680 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.492729 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.492740 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.492760 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.492773 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:00Z","lastTransitionTime":"2025-10-01T09:37:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.523485 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.523575 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.523633 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:37:00 crc kubenswrapper[4787]: E1001 09:37:00.523718 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:37:00 crc kubenswrapper[4787]: E1001 09:37:00.523866 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:37:00 crc kubenswrapper[4787]: E1001 09:37:00.524054 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.595592 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.595653 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.595671 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.595695 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.595712 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:00Z","lastTransitionTime":"2025-10-01T09:37:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.699233 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.699324 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.699342 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.699367 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.699390 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:00Z","lastTransitionTime":"2025-10-01T09:37:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.803401 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.803458 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.803474 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.803496 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.803510 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:00Z","lastTransitionTime":"2025-10-01T09:37:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.906837 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.906896 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.906908 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.906930 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:00 crc kubenswrapper[4787]: I1001 09:37:00.906943 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:00Z","lastTransitionTime":"2025-10-01T09:37:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.010783 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.010879 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.010906 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.010942 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.010966 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:01Z","lastTransitionTime":"2025-10-01T09:37:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.114201 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.114268 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.114280 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.114303 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.114316 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:01Z","lastTransitionTime":"2025-10-01T09:37:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.216637 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.216680 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.216690 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.216705 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.216717 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:01Z","lastTransitionTime":"2025-10-01T09:37:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.318975 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.319020 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.319035 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.319052 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.319065 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:01Z","lastTransitionTime":"2025-10-01T09:37:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.422153 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.422202 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.422218 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.422235 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.422246 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:01Z","lastTransitionTime":"2025-10-01T09:37:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.522895 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:37:01 crc kubenswrapper[4787]: E1001 09:37:01.523106 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.526657 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.526710 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.526723 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.526755 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.526772 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:01Z","lastTransitionTime":"2025-10-01T09:37:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.630004 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.630052 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.630068 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.630123 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.630148 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:01Z","lastTransitionTime":"2025-10-01T09:37:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.732777 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.732822 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.732831 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.732846 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.732855 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:01Z","lastTransitionTime":"2025-10-01T09:37:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.835868 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.835937 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.835949 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.835964 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.835975 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:01Z","lastTransitionTime":"2025-10-01T09:37:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.938949 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.939006 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.939019 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.939032 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:01 crc kubenswrapper[4787]: I1001 09:37:01.939041 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:01Z","lastTransitionTime":"2025-10-01T09:37:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.040640 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.040667 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.040675 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.040688 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.040698 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:02Z","lastTransitionTime":"2025-10-01T09:37:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.143661 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.143728 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.143741 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.143758 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.143771 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:02Z","lastTransitionTime":"2025-10-01T09:37:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.246642 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.246698 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.246710 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.246727 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.246739 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:02Z","lastTransitionTime":"2025-10-01T09:37:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.349617 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.349671 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.349693 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.349708 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.349719 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:02Z","lastTransitionTime":"2025-10-01T09:37:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.452032 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.452089 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.452100 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.452122 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.452135 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:02Z","lastTransitionTime":"2025-10-01T09:37:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.523799 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:37:02 crc kubenswrapper[4787]: E1001 09:37:02.524298 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.525538 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:37:02 crc kubenswrapper[4787]: E1001 09:37:02.525620 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.525856 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:37:02 crc kubenswrapper[4787]: E1001 09:37:02.525919 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.554727 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.554782 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.554793 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.554811 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.554822 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:02Z","lastTransitionTime":"2025-10-01T09:37:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.657118 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.657156 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.657193 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.657210 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.657219 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:02Z","lastTransitionTime":"2025-10-01T09:37:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.759307 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.759341 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.759354 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.759371 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.759382 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:02Z","lastTransitionTime":"2025-10-01T09:37:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.861130 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.861167 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.861175 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.861188 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.861197 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:02Z","lastTransitionTime":"2025-10-01T09:37:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.963172 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.963204 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.963212 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.963225 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:02 crc kubenswrapper[4787]: I1001 09:37:02.963235 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:02Z","lastTransitionTime":"2025-10-01T09:37:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.065054 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.065287 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.065297 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.065309 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.065317 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:03Z","lastTransitionTime":"2025-10-01T09:37:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.167586 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.167641 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.167657 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.167679 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.167691 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:03Z","lastTransitionTime":"2025-10-01T09:37:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.270421 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.270467 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.270479 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.270498 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.270514 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:03Z","lastTransitionTime":"2025-10-01T09:37:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.372565 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.372618 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.372629 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.372644 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.372655 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:03Z","lastTransitionTime":"2025-10-01T09:37:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.475315 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.475386 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.475394 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.475408 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.475417 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:03Z","lastTransitionTime":"2025-10-01T09:37:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.523543 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:37:03 crc kubenswrapper[4787]: E1001 09:37:03.523670 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.577833 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.577894 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.577907 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.577922 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.577933 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:03Z","lastTransitionTime":"2025-10-01T09:37:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.680132 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.680170 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.680181 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.680194 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.680202 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:03Z","lastTransitionTime":"2025-10-01T09:37:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.782068 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.782140 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.782152 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.782166 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.782177 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:03Z","lastTransitionTime":"2025-10-01T09:37:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.884690 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.884735 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.884746 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.884761 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.884774 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:03Z","lastTransitionTime":"2025-10-01T09:37:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.987546 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.987575 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.987583 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.987596 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:03 crc kubenswrapper[4787]: I1001 09:37:03.987604 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:03Z","lastTransitionTime":"2025-10-01T09:37:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.090058 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.090112 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.090121 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.090134 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.090144 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:04Z","lastTransitionTime":"2025-10-01T09:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.192109 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.192149 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.192159 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.192175 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.192184 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:04Z","lastTransitionTime":"2025-10-01T09:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.294276 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.294314 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.294326 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.294341 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.294351 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:04Z","lastTransitionTime":"2025-10-01T09:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.396213 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.396269 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.396279 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.396295 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.396307 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:04Z","lastTransitionTime":"2025-10-01T09:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.498620 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.498858 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.498962 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.499047 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.499186 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:04Z","lastTransitionTime":"2025-10-01T09:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.523311 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:37:04 crc kubenswrapper[4787]: E1001 09:37:04.523686 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.523434 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:37:04 crc kubenswrapper[4787]: E1001 09:37:04.524140 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.523371 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:37:04 crc kubenswrapper[4787]: E1001 09:37:04.524389 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.601742 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.601777 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.601788 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.601802 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.601812 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:04Z","lastTransitionTime":"2025-10-01T09:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.704454 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.704721 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.704792 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.704855 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.704916 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:04Z","lastTransitionTime":"2025-10-01T09:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.807692 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.808027 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.808061 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.808094 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.808113 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:04Z","lastTransitionTime":"2025-10-01T09:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.874445 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.874477 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.874487 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.874499 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.874507 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:04Z","lastTransitionTime":"2025-10-01T09:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:04 crc kubenswrapper[4787]: E1001 09:37:04.888173 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.892229 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.892442 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.892508 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.892593 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.892677 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:04Z","lastTransitionTime":"2025-10-01T09:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:04 crc kubenswrapper[4787]: E1001 09:37:04.908911 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.913927 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.914070 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.914221 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.914309 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.914369 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:04Z","lastTransitionTime":"2025-10-01T09:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:04 crc kubenswrapper[4787]: E1001 09:37:04.932766 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.936778 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.936937 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.937012 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.937100 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.937200 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:04Z","lastTransitionTime":"2025-10-01T09:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:04 crc kubenswrapper[4787]: E1001 09:37:04.949442 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.955414 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.955464 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.955477 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.955495 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.955509 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:04Z","lastTransitionTime":"2025-10-01T09:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:04 crc kubenswrapper[4787]: E1001 09:37:04.969679 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:04Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:04 crc kubenswrapper[4787]: E1001 09:37:04.969847 4787 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.971415 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.971475 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.971488 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.971509 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:04 crc kubenswrapper[4787]: I1001 09:37:04.971521 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:04Z","lastTransitionTime":"2025-10-01T09:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.073965 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.074018 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.074036 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.074055 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.074067 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:05Z","lastTransitionTime":"2025-10-01T09:37:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.177250 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.177297 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.177310 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.177325 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.177334 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:05Z","lastTransitionTime":"2025-10-01T09:37:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.280194 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.280240 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.280255 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.280273 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.280286 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:05Z","lastTransitionTime":"2025-10-01T09:37:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.383607 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.383669 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.383678 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.383692 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.383702 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:05Z","lastTransitionTime":"2025-10-01T09:37:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.486686 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.486739 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.486751 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.486767 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.486778 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:05Z","lastTransitionTime":"2025-10-01T09:37:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.523742 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:37:05 crc kubenswrapper[4787]: E1001 09:37:05.523949 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.589476 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.589524 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.589535 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.589552 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.589563 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:05Z","lastTransitionTime":"2025-10-01T09:37:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.691787 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.691828 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.691838 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.691850 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.691860 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:05Z","lastTransitionTime":"2025-10-01T09:37:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.794121 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.794147 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.794155 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.794169 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.794178 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:05Z","lastTransitionTime":"2025-10-01T09:37:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.897098 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.897158 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.897172 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.897194 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:05 crc kubenswrapper[4787]: I1001 09:37:05.897207 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:05Z","lastTransitionTime":"2025-10-01T09:37:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.000629 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.000685 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.000701 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.000720 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.000731 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:06Z","lastTransitionTime":"2025-10-01T09:37:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.104269 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.104329 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.104340 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.104367 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.104380 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:06Z","lastTransitionTime":"2025-10-01T09:37:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.207567 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.207618 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.207638 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.207661 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.207674 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:06Z","lastTransitionTime":"2025-10-01T09:37:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.310828 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.310864 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.310873 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.310889 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.310899 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:06Z","lastTransitionTime":"2025-10-01T09:37:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.413237 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.413288 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.413296 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.413310 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.413322 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:06Z","lastTransitionTime":"2025-10-01T09:37:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.516239 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.516308 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.516332 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.516360 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.516380 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:06Z","lastTransitionTime":"2025-10-01T09:37:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.523506 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.523575 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:37:06 crc kubenswrapper[4787]: E1001 09:37:06.523624 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:37:06 crc kubenswrapper[4787]: E1001 09:37:06.523727 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.523807 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:37:06 crc kubenswrapper[4787]: E1001 09:37:06.523962 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.618607 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.618647 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.618657 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.618669 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.618678 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:06Z","lastTransitionTime":"2025-10-01T09:37:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.722578 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.722661 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.722680 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.722710 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.722731 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:06Z","lastTransitionTime":"2025-10-01T09:37:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.824793 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.824839 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.824847 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.824861 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.824871 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:06Z","lastTransitionTime":"2025-10-01T09:37:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.927414 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.927456 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.927466 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.927481 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:06 crc kubenswrapper[4787]: I1001 09:37:06.927490 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:06Z","lastTransitionTime":"2025-10-01T09:37:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.030898 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.030941 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.030954 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.030969 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.030980 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:07Z","lastTransitionTime":"2025-10-01T09:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.133575 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.133650 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.133666 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.133680 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.133691 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:07Z","lastTransitionTime":"2025-10-01T09:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.236413 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.236451 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.236460 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.236472 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.236481 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:07Z","lastTransitionTime":"2025-10-01T09:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.339057 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.339130 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.339144 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.339160 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.339172 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:07Z","lastTransitionTime":"2025-10-01T09:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.441230 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.441266 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.441278 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.441294 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.441306 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:07Z","lastTransitionTime":"2025-10-01T09:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.523327 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:37:07 crc kubenswrapper[4787]: E1001 09:37:07.523450 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.544211 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.544245 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.544253 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.544266 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.544277 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:07Z","lastTransitionTime":"2025-10-01T09:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.646771 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.646869 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.646884 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.646899 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.646910 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:07Z","lastTransitionTime":"2025-10-01T09:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.749853 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.749916 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.749934 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.749960 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.749975 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:07Z","lastTransitionTime":"2025-10-01T09:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.852500 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.852550 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.852561 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.852578 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.852589 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:07Z","lastTransitionTime":"2025-10-01T09:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.955609 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.955648 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.955657 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.955679 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:07 crc kubenswrapper[4787]: I1001 09:37:07.955691 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:07Z","lastTransitionTime":"2025-10-01T09:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.058531 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.058580 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.058592 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.058611 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.058625 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:08Z","lastTransitionTime":"2025-10-01T09:37:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.160819 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.160886 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.160908 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.160937 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.160960 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:08Z","lastTransitionTime":"2025-10-01T09:37:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.263587 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.263629 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.263638 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.263653 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.263662 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:08Z","lastTransitionTime":"2025-10-01T09:37:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.365262 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.365295 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.365304 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.365340 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.365353 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:08Z","lastTransitionTime":"2025-10-01T09:37:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.466918 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.466953 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.466964 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.466977 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.466988 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:08Z","lastTransitionTime":"2025-10-01T09:37:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.522835 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.522877 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.523429 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:37:08 crc kubenswrapper[4787]: E1001 09:37:08.523636 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:37:08 crc kubenswrapper[4787]: E1001 09:37:08.523767 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:37:08 crc kubenswrapper[4787]: E1001 09:37:08.523900 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.524048 4787 scope.go:117] "RemoveContainer" containerID="e77339fe976956bb53dcd610364af111550cb6bfa1103daf7f706a9ee74c4439" Oct 01 09:37:08 crc kubenswrapper[4787]: E1001 09:37:08.524423 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dmsxm_openshift-ovn-kubernetes(eadf2545-886e-4642-b909-704c6cd4134e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" podUID="eadf2545-886e-4642-b909-704c6cd4134e" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.535893 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:08Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.547014 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4990da342a20f6ab1e97774294bd6856c49ca17de21d893a638c13c8e727dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:08Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.564150 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b093e96a244fdec4c5cdce62650d4da31c2ec46207a9c08bbd0a28bf286abb52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4664cb1dde25ada88989e81f0319aedc6dbd363f608a50623efaa3eeb8d00763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:08Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.569096 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.569129 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.569140 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.569155 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.569181 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:08Z","lastTransitionTime":"2025-10-01T09:37:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.577847 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d95f8b1aac879d1746d8a8af7e1d5e672b36c31d5b95c23fee70b237bc3716e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:08Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.588140 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7aa5c938897ea621fc7c71c86e6a1dd2acca273e13afe71e9a27fc81f34cf2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:08Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.603938 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c68b374f72c2a689eb09480e7ae3e88691daf0cb380a52f655eb1c651a66a281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:08Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.622791 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e77339fe976956bb53dcd610364af111550cb6bfa1103daf7f706a9ee74c4439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e77339fe976956bb53dcd610364af111550cb6bfa1103daf7f706a9ee74c4439\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:36:56Z\\\",\\\"message\\\":\\\"{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: kube-storage-version-migrator-operator,},ClusterIP:10.217.5.36,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.36],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1001 09:36:56.575330 6417 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dmsxm_openshift-ovn-kubernetes(eadf2545-886e-4642-b909-704c6cd4134e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:08Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.635185 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6zz6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb4a902-9efa-4088-a236-9a634fd085fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k5bxp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k5bxp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6zz6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:08Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.650269 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0814bed7-1117-4dff-97ef-a3cd1240e615\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c91542e7254c8289b734b3ffa8e8226f2445c915c264e09c179b99604abad1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:36:27Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1001 09:36:22.478145 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 09:36:22.478910 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3375976971/tls.crt::/tmp/serving-cert-3375976971/tls.key\\\\\\\"\\\\nI1001 09:36:27.943129 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 09:36:27.955126 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 09:36:27.955153 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 09:36:27.955179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 09:36:27.955185 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 09:36:27.963955 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1001 09:36:27.963979 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1001 09:36:27.963986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964008 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 09:36:27.964015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 09:36:27.964018 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 09:36:27.964022 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1001 09:36:27.966112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:08Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.663142 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc37698-af76-4818-b216-ffe35af67929\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1bb16edb25b63e957cc0bceb04a9fa9244e68881cb45be7c7d627c5faa74232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://995030421e0cbfeb7baa8ccb574d269dd178f8185bcaa9000fe5e7fc221b1188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d4c78dd607b81a28bcfe0ffa69c02628cf06aff5fd6a574069639c5bac044a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://144123d273f3f1af4fb857d731a16d9891b097c98229a035eb132f360fa8b2f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:08Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.670922 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.670951 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.670960 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.670974 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.670984 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:08Z","lastTransitionTime":"2025-10-01T09:37:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.675875 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72701c6061105b7bb81545b5f0b035b220b22b2f8009e82809a037870c625685\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:08Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.689452 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:08Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.701885 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:08Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.714465 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"180a8b2f-41ef-460c-98f2-2914e3b8f9d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f445035e4b6faf9fca8e0dbe8c69668de8825ccebd6ace48f80685c88312b707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f63e01c8f45be048cd6ade60cb51f054fbea42bb51d7b5cd32f71cf1f15ce21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1d95bbc5f966defa47192b7cc53c32f75894f4b51ce8411d7bf828e04c282b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c3cd7c1e19e00dbc9cb8cb8f3f857ae636ff3818c589769f15aa87cb94327ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c3cd7c1e19e00dbc9cb8cb8f3f857ae636ff3818c589769f15aa87cb94327ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:08Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.725109 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebc8a6e2b9c6104b0f773e99d68a62e1b2bf0adc05e6f291cf91e7a8cf1bc121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:08Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.735682 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"049825a6-0cec-4adf-8b77-4889bdc4d56f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18044055eb1f6cfd7cb05f330314231b1c563c34a51ba5ed79073988eb563cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc8192c84986df76c9a45deba61aaaa1b0a019fce0000e1c9916c4fe31d1ba1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-84q9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:08Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.746389 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:08Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.773179 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.773237 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.773256 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.773281 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.773302 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:08Z","lastTransitionTime":"2025-10-01T09:37:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.876035 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.876277 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.876339 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.876405 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.876463 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:08Z","lastTransitionTime":"2025-10-01T09:37:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.978616 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.978916 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.979056 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.979184 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:08 crc kubenswrapper[4787]: I1001 09:37:08.979285 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:08Z","lastTransitionTime":"2025-10-01T09:37:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.082146 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.082619 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.082851 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.083053 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.083262 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:09Z","lastTransitionTime":"2025-10-01T09:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.185431 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.185467 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.185476 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.185529 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.185545 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:09Z","lastTransitionTime":"2025-10-01T09:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.288621 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.288659 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.288670 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.288687 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.288702 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:09Z","lastTransitionTime":"2025-10-01T09:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.391812 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.391842 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.391856 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.391872 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.391884 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:09Z","lastTransitionTime":"2025-10-01T09:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.494135 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.494436 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.494522 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.494641 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.494743 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:09Z","lastTransitionTime":"2025-10-01T09:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.522879 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:37:09 crc kubenswrapper[4787]: E1001 09:37:09.523044 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.597095 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.597151 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.597162 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.597180 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.597192 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:09Z","lastTransitionTime":"2025-10-01T09:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.699833 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.699887 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.699903 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.699922 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.699934 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:09Z","lastTransitionTime":"2025-10-01T09:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.801991 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.802129 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.802145 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.802166 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.802184 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:09Z","lastTransitionTime":"2025-10-01T09:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.904308 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.904341 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.904350 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.904362 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:09 crc kubenswrapper[4787]: I1001 09:37:09.904371 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:09Z","lastTransitionTime":"2025-10-01T09:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.007161 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.007199 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.007207 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.007221 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.007232 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:10Z","lastTransitionTime":"2025-10-01T09:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.109424 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.109544 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.109588 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.109623 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.109642 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:10Z","lastTransitionTime":"2025-10-01T09:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.212993 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.213043 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.213053 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.213068 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.213099 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:10Z","lastTransitionTime":"2025-10-01T09:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.315575 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.315653 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.315663 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.315677 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.315687 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:10Z","lastTransitionTime":"2025-10-01T09:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.417440 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.417494 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.417508 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.417525 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.417536 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:10Z","lastTransitionTime":"2025-10-01T09:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.519928 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.519960 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.519969 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.519984 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.519995 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:10Z","lastTransitionTime":"2025-10-01T09:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.523249 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.523311 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:37:10 crc kubenswrapper[4787]: E1001 09:37:10.523338 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.523451 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:37:10 crc kubenswrapper[4787]: E1001 09:37:10.523506 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:37:10 crc kubenswrapper[4787]: E1001 09:37:10.523442 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.622552 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.622585 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.622595 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.622608 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.622616 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:10Z","lastTransitionTime":"2025-10-01T09:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.725476 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.725562 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.725580 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.725603 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.725622 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:10Z","lastTransitionTime":"2025-10-01T09:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.828226 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.828480 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.828547 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.828625 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.828704 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:10Z","lastTransitionTime":"2025-10-01T09:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.930674 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.930725 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.930738 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.930757 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:10 crc kubenswrapper[4787]: I1001 09:37:10.930771 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:10Z","lastTransitionTime":"2025-10-01T09:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.033545 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.033580 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.033588 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.033630 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.033643 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:11Z","lastTransitionTime":"2025-10-01T09:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.136009 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.136041 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.136048 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.136061 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.136070 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:11Z","lastTransitionTime":"2025-10-01T09:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.238683 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.238738 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.238763 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.238776 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.238786 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:11Z","lastTransitionTime":"2025-10-01T09:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.341975 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.342015 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.342024 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.342040 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.342049 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:11Z","lastTransitionTime":"2025-10-01T09:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.444642 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.444692 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.444708 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.444731 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.444748 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:11Z","lastTransitionTime":"2025-10-01T09:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.523798 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:37:11 crc kubenswrapper[4787]: E1001 09:37:11.524327 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.547321 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.547371 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.547383 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.547403 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.547415 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:11Z","lastTransitionTime":"2025-10-01T09:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.650128 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.650199 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.650219 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.650252 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.650273 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:11Z","lastTransitionTime":"2025-10-01T09:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.752321 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.752353 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.752365 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.752379 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.752388 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:11Z","lastTransitionTime":"2025-10-01T09:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.855257 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.855579 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.855671 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.855766 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.856144 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:11Z","lastTransitionTime":"2025-10-01T09:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.958494 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.958558 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.958571 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.958592 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:11 crc kubenswrapper[4787]: I1001 09:37:11.958604 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:11Z","lastTransitionTime":"2025-10-01T09:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.060578 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.060621 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.060633 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.060649 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.060659 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:12Z","lastTransitionTime":"2025-10-01T09:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.163256 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.163300 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.163309 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.163323 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.163334 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:12Z","lastTransitionTime":"2025-10-01T09:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.265908 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.265946 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.265954 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.265967 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.265976 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:12Z","lastTransitionTime":"2025-10-01T09:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.368699 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.368756 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.368769 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.368788 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.368801 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:12Z","lastTransitionTime":"2025-10-01T09:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.470646 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.470692 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.470704 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.470721 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.470734 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:12Z","lastTransitionTime":"2025-10-01T09:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.523514 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.523564 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.523590 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:37:12 crc kubenswrapper[4787]: E1001 09:37:12.523664 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:37:12 crc kubenswrapper[4787]: E1001 09:37:12.523720 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:37:12 crc kubenswrapper[4787]: E1001 09:37:12.523847 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.573127 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.573164 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.573175 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.573188 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.573196 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:12Z","lastTransitionTime":"2025-10-01T09:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.675745 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.675789 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.675800 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.675816 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.675827 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:12Z","lastTransitionTime":"2025-10-01T09:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.778208 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.778250 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.778259 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.778273 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.778282 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:12Z","lastTransitionTime":"2025-10-01T09:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.885465 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.885513 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.885525 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.885546 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.885559 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:12Z","lastTransitionTime":"2025-10-01T09:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.988140 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.988262 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.988279 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.988316 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:12 crc kubenswrapper[4787]: I1001 09:37:12.988333 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:12Z","lastTransitionTime":"2025-10-01T09:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.090472 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.090525 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.090542 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.090565 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.090583 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:13Z","lastTransitionTime":"2025-10-01T09:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.192492 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.192543 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.192552 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.192566 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.192575 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:13Z","lastTransitionTime":"2025-10-01T09:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.295432 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.295486 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.295507 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.295526 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.295538 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:13Z","lastTransitionTime":"2025-10-01T09:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.398253 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.398290 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.398299 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.398312 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.398324 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:13Z","lastTransitionTime":"2025-10-01T09:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.500726 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.500762 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.500774 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.500792 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.500805 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:13Z","lastTransitionTime":"2025-10-01T09:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.523031 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:37:13 crc kubenswrapper[4787]: E1001 09:37:13.523138 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.604259 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.604343 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.604370 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.604417 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.604445 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:13Z","lastTransitionTime":"2025-10-01T09:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.706726 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.706772 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.706781 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.706797 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.706806 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:13Z","lastTransitionTime":"2025-10-01T09:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.808833 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.808868 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.808878 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.808891 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.808901 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:13Z","lastTransitionTime":"2025-10-01T09:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.910694 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.910741 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.910751 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.910767 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:13 crc kubenswrapper[4787]: I1001 09:37:13.910777 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:13Z","lastTransitionTime":"2025-10-01T09:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.013739 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.013799 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.013811 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.013829 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.013842 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:14Z","lastTransitionTime":"2025-10-01T09:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.115873 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.115927 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.115935 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.115947 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.115956 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:14Z","lastTransitionTime":"2025-10-01T09:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.219490 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.219582 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.219613 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.219681 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.219707 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:14Z","lastTransitionTime":"2025-10-01T09:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.322328 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.322400 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.322428 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.322457 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.322481 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:14Z","lastTransitionTime":"2025-10-01T09:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.329273 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0eb4a902-9efa-4088-a236-9a634fd085fd-metrics-certs\") pod \"network-metrics-daemon-6zz6g\" (UID: \"0eb4a902-9efa-4088-a236-9a634fd085fd\") " pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:37:14 crc kubenswrapper[4787]: E1001 09:37:14.329425 4787 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 09:37:14 crc kubenswrapper[4787]: E1001 09:37:14.329485 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0eb4a902-9efa-4088-a236-9a634fd085fd-metrics-certs podName:0eb4a902-9efa-4088-a236-9a634fd085fd nodeName:}" failed. No retries permitted until 2025-10-01 09:37:46.3294677 +0000 UTC m=+98.444611857 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0eb4a902-9efa-4088-a236-9a634fd085fd-metrics-certs") pod "network-metrics-daemon-6zz6g" (UID: "0eb4a902-9efa-4088-a236-9a634fd085fd") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.425310 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.425368 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.425387 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.425403 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.425436 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:14Z","lastTransitionTime":"2025-10-01T09:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.523371 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.523371 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:37:14 crc kubenswrapper[4787]: E1001 09:37:14.523650 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.523369 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:37:14 crc kubenswrapper[4787]: E1001 09:37:14.523757 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:37:14 crc kubenswrapper[4787]: E1001 09:37:14.523508 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.527702 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.527774 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.527786 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.527832 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.527845 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:14Z","lastTransitionTime":"2025-10-01T09:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.630232 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.630278 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.630289 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.630304 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.630316 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:14Z","lastTransitionTime":"2025-10-01T09:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.732204 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.732233 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.732241 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.732253 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.732262 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:14Z","lastTransitionTime":"2025-10-01T09:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.835058 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.835113 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.835124 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.835141 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.835150 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:14Z","lastTransitionTime":"2025-10-01T09:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.937231 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.937296 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.937305 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.937320 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:14 crc kubenswrapper[4787]: I1001 09:37:14.937338 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:14Z","lastTransitionTime":"2025-10-01T09:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.039834 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.039871 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.039879 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.039894 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.039902 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:15Z","lastTransitionTime":"2025-10-01T09:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.142126 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.142156 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.142183 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.142197 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.142206 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:15Z","lastTransitionTime":"2025-10-01T09:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.166776 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.166818 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.166831 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.166848 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.166866 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:15Z","lastTransitionTime":"2025-10-01T09:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:15 crc kubenswrapper[4787]: E1001 09:37:15.181275 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:15Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.184091 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.184120 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.184129 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.184143 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.184153 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:15Z","lastTransitionTime":"2025-10-01T09:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:15 crc kubenswrapper[4787]: E1001 09:37:15.194850 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:15Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.198045 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.198086 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.198095 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.198122 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.198142 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:15Z","lastTransitionTime":"2025-10-01T09:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:15 crc kubenswrapper[4787]: E1001 09:37:15.208629 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:15Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.211892 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.211925 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.211934 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.211949 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.211961 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:15Z","lastTransitionTime":"2025-10-01T09:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:15 crc kubenswrapper[4787]: E1001 09:37:15.223204 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:15Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.226537 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.226568 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.226577 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.226590 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.226600 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:15Z","lastTransitionTime":"2025-10-01T09:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:15 crc kubenswrapper[4787]: E1001 09:37:15.238947 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:15Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:15 crc kubenswrapper[4787]: E1001 09:37:15.239049 4787 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.244347 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.244374 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.244383 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.244396 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.244406 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:15Z","lastTransitionTime":"2025-10-01T09:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.346019 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.346046 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.346054 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.346065 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.346086 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:15Z","lastTransitionTime":"2025-10-01T09:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.448224 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.448257 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.448266 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.448280 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.448289 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:15Z","lastTransitionTime":"2025-10-01T09:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.523193 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:37:15 crc kubenswrapper[4787]: E1001 09:37:15.523379 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.551236 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.551276 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.551285 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.551301 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.551310 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:15Z","lastTransitionTime":"2025-10-01T09:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.654772 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.654814 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.654824 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.654843 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.654853 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:15Z","lastTransitionTime":"2025-10-01T09:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.757420 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.757463 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.757475 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.757493 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.757510 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:15Z","lastTransitionTime":"2025-10-01T09:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.859031 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.859087 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.859097 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.859112 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.859123 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:15Z","lastTransitionTime":"2025-10-01T09:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.908830 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z8tjz_53f5ffd2-7649-4c9e-929f-a68187409644/kube-multus/0.log" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.909107 4787 generic.go:334] "Generic (PLEG): container finished" podID="53f5ffd2-7649-4c9e-929f-a68187409644" containerID="cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885" exitCode=1 Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.909219 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-z8tjz" event={"ID":"53f5ffd2-7649-4c9e-929f-a68187409644","Type":"ContainerDied","Data":"cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885"} Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.909752 4787 scope.go:117] "RemoveContainer" containerID="cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.922669 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0814bed7-1117-4dff-97ef-a3cd1240e615\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c91542e7254c8289b734b3ffa8e8226f2445c915c264e09c179b99604abad1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:36:27Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1001 09:36:22.478145 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 09:36:22.478910 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3375976971/tls.crt::/tmp/serving-cert-3375976971/tls.key\\\\\\\"\\\\nI1001 09:36:27.943129 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 09:36:27.955126 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 09:36:27.955153 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 09:36:27.955179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 09:36:27.955185 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 09:36:27.963955 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1001 09:36:27.963979 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1001 09:36:27.963986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964008 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 09:36:27.964015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 09:36:27.964018 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 09:36:27.964022 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1001 09:36:27.966112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:15Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.934634 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7aa5c938897ea621fc7c71c86e6a1dd2acca273e13afe71e9a27fc81f34cf2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:15Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.952205 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c68b374f72c2a689eb09480e7ae3e88691daf0cb380a52f655eb1c651a66a281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:15Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.963636 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.963683 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.963693 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.963709 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.963719 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:15Z","lastTransitionTime":"2025-10-01T09:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.970471 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e77339fe976956bb53dcd610364af111550cb6bfa1103daf7f706a9ee74c4439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e77339fe976956bb53dcd610364af111550cb6bfa1103daf7f706a9ee74c4439\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:36:56Z\\\",\\\"message\\\":\\\"{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: kube-storage-version-migrator-operator,},ClusterIP:10.217.5.36,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.36],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1001 09:36:56.575330 6417 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dmsxm_openshift-ovn-kubernetes(eadf2545-886e-4642-b909-704c6cd4134e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:15Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.982250 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6zz6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb4a902-9efa-4088-a236-9a634fd085fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k5bxp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k5bxp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6zz6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:15Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:15 crc kubenswrapper[4787]: I1001 09:37:15.993323 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"180a8b2f-41ef-460c-98f2-2914e3b8f9d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f445035e4b6faf9fca8e0dbe8c69668de8825ccebd6ace48f80685c88312b707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f63e01c8f45be048cd6ade60cb51f054fbea42bb51d7b5cd32f71cf1f15ce21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1d95bbc5f966defa47192b7cc53c32f75894f4b51ce8411d7bf828e04c282b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c3cd7c1e19e00dbc9cb8cb8f3f857ae636ff3818c589769f15aa87cb94327ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c3cd7c1e19e00dbc9cb8cb8f3f857ae636ff3818c589769f15aa87cb94327ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:15Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.005766 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc37698-af76-4818-b216-ffe35af67929\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1bb16edb25b63e957cc0bceb04a9fa9244e68881cb45be7c7d627c5faa74232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://995030421e0cbfeb7baa8ccb574d269dd178f8185bcaa9000fe5e7fc221b1188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d4c78dd607b81a28bcfe0ffa69c02628cf06aff5fd6a574069639c5bac044a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://144123d273f3f1af4fb857d731a16d9891b097c98229a035eb132f360fa8b2f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:16Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.017998 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72701c6061105b7bb81545b5f0b035b220b22b2f8009e82809a037870c625685\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:16Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.032374 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:16Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.047925 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"message\\\":\\\"2025-10-01T09:36:30+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_54bdebc1-b80c-438d-921a-17ba60e91a86\\\\n2025-10-01T09:36:30+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_54bdebc1-b80c-438d-921a-17ba60e91a86 to /host/opt/cni/bin/\\\\n2025-10-01T09:36:30Z [verbose] multus-daemon started\\\\n2025-10-01T09:36:30Z [verbose] Readiness Indicator file check\\\\n2025-10-01T09:37:15Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:16Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.059745 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:16Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.075415 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebc8a6e2b9c6104b0f773e99d68a62e1b2bf0adc05e6f291cf91e7a8cf1bc121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:16Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.077086 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.077124 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.077133 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.077147 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.077156 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:16Z","lastTransitionTime":"2025-10-01T09:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.088449 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"049825a6-0cec-4adf-8b77-4889bdc4d56f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18044055eb1f6cfd7cb05f330314231b1c563c34a51ba5ed79073988eb563cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc8192c84986df76c9a45deba61aaaa1b0a019fce0000e1c9916c4fe31d1ba1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-84q9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:16Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.099617 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:16Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.109659 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4990da342a20f6ab1e97774294bd6856c49ca17de21d893a638c13c8e727dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:16Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.120737 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b093e96a244fdec4c5cdce62650d4da31c2ec46207a9c08bbd0a28bf286abb52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4664cb1dde25ada88989e81f0319aedc6dbd363f608a50623efaa3eeb8d00763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:16Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.132656 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d95f8b1aac879d1746d8a8af7e1d5e672b36c31d5b95c23fee70b237bc3716e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:16Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.179860 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.179905 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.179918 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.179935 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.179945 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:16Z","lastTransitionTime":"2025-10-01T09:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.282862 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.282912 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.282929 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.282957 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.282994 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:16Z","lastTransitionTime":"2025-10-01T09:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.384841 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.384880 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.384892 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.384909 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.384923 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:16Z","lastTransitionTime":"2025-10-01T09:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.487893 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.487931 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.487940 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.487956 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.487970 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:16Z","lastTransitionTime":"2025-10-01T09:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.523554 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.523570 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:37:16 crc kubenswrapper[4787]: E1001 09:37:16.523810 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:37:16 crc kubenswrapper[4787]: E1001 09:37:16.523924 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.524219 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:37:16 crc kubenswrapper[4787]: E1001 09:37:16.524317 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.590787 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.590833 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.590845 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.590865 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.590877 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:16Z","lastTransitionTime":"2025-10-01T09:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.693296 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.693362 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.693370 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.693382 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.693391 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:16Z","lastTransitionTime":"2025-10-01T09:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.795609 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.795651 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.795663 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.795678 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.795689 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:16Z","lastTransitionTime":"2025-10-01T09:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.897708 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.897749 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.897760 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.897776 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.897787 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:16Z","lastTransitionTime":"2025-10-01T09:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.914638 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z8tjz_53f5ffd2-7649-4c9e-929f-a68187409644/kube-multus/0.log" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.914707 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-z8tjz" event={"ID":"53f5ffd2-7649-4c9e-929f-a68187409644","Type":"ContainerStarted","Data":"8c830596f83deb4fa998fa9af881c7eb1f3673588f444edc9cbf895c19b8248c"} Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.934360 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:16Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.965691 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4990da342a20f6ab1e97774294bd6856c49ca17de21d893a638c13c8e727dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:16Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:16 crc kubenswrapper[4787]: I1001 09:37:16.988556 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b093e96a244fdec4c5cdce62650d4da31c2ec46207a9c08bbd0a28bf286abb52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4664cb1dde25ada88989e81f0319aedc6dbd363f608a50623efaa3eeb8d00763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:16Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.000584 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.000620 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.000630 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.000646 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.000660 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:17Z","lastTransitionTime":"2025-10-01T09:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.026558 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d95f8b1aac879d1746d8a8af7e1d5e672b36c31d5b95c23fee70b237bc3716e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:17Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.042971 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c68b374f72c2a689eb09480e7ae3e88691daf0cb380a52f655eb1c651a66a281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:17Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.063025 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e77339fe976956bb53dcd610364af111550cb6bfa1103daf7f706a9ee74c4439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e77339fe976956bb53dcd610364af111550cb6bfa1103daf7f706a9ee74c4439\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:36:56Z\\\",\\\"message\\\":\\\"{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: kube-storage-version-migrator-operator,},ClusterIP:10.217.5.36,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.36],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1001 09:36:56.575330 6417 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dmsxm_openshift-ovn-kubernetes(eadf2545-886e-4642-b909-704c6cd4134e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:17Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.074878 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6zz6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb4a902-9efa-4088-a236-9a634fd085fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k5bxp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k5bxp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6zz6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:17Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.087296 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0814bed7-1117-4dff-97ef-a3cd1240e615\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c91542e7254c8289b734b3ffa8e8226f2445c915c264e09c179b99604abad1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:36:27Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1001 09:36:22.478145 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 09:36:22.478910 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3375976971/tls.crt::/tmp/serving-cert-3375976971/tls.key\\\\\\\"\\\\nI1001 09:36:27.943129 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 09:36:27.955126 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 09:36:27.955153 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 09:36:27.955179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 09:36:27.955185 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 09:36:27.963955 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1001 09:36:27.963979 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1001 09:36:27.963986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964008 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 09:36:27.964015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 09:36:27.964018 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 09:36:27.964022 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1001 09:36:27.966112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:17Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.098270 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7aa5c938897ea621fc7c71c86e6a1dd2acca273e13afe71e9a27fc81f34cf2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:17Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.102520 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.102562 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.102574 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.102590 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.102602 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:17Z","lastTransitionTime":"2025-10-01T09:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.112098 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72701c6061105b7bb81545b5f0b035b220b22b2f8009e82809a037870c625685\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:17Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.125539 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:17Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.139789 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c830596f83deb4fa998fa9af881c7eb1f3673588f444edc9cbf895c19b8248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"message\\\":\\\"2025-10-01T09:36:30+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_54bdebc1-b80c-438d-921a-17ba60e91a86\\\\n2025-10-01T09:36:30+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_54bdebc1-b80c-438d-921a-17ba60e91a86 to /host/opt/cni/bin/\\\\n2025-10-01T09:36:30Z [verbose] multus-daemon started\\\\n2025-10-01T09:36:30Z [verbose] Readiness Indicator file check\\\\n2025-10-01T09:37:15Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:17Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.152427 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"180a8b2f-41ef-460c-98f2-2914e3b8f9d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f445035e4b6faf9fca8e0dbe8c69668de8825ccebd6ace48f80685c88312b707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f63e01c8f45be048cd6ade60cb51f054fbea42bb51d7b5cd32f71cf1f15ce21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1d95bbc5f966defa47192b7cc53c32f75894f4b51ce8411d7bf828e04c282b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c3cd7c1e19e00dbc9cb8cb8f3f857ae636ff3818c589769f15aa87cb94327ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c3cd7c1e19e00dbc9cb8cb8f3f857ae636ff3818c589769f15aa87cb94327ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:17Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.166987 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc37698-af76-4818-b216-ffe35af67929\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1bb16edb25b63e957cc0bceb04a9fa9244e68881cb45be7c7d627c5faa74232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://995030421e0cbfeb7baa8ccb574d269dd178f8185bcaa9000fe5e7fc221b1188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d4c78dd607b81a28bcfe0ffa69c02628cf06aff5fd6a574069639c5bac044a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://144123d273f3f1af4fb857d731a16d9891b097c98229a035eb132f360fa8b2f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:17Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.179593 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"049825a6-0cec-4adf-8b77-4889bdc4d56f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18044055eb1f6cfd7cb05f330314231b1c563c34a51ba5ed79073988eb563cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc8192c84986df76c9a45deba61aaaa1b0a019fce0000e1c9916c4fe31d1ba1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-84q9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:17Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.192116 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:17Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.203373 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebc8a6e2b9c6104b0f773e99d68a62e1b2bf0adc05e6f291cf91e7a8cf1bc121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:17Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.205395 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.205450 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.205463 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.205480 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.205835 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:17Z","lastTransitionTime":"2025-10-01T09:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.309562 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.309605 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.309616 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.309634 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.309645 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:17Z","lastTransitionTime":"2025-10-01T09:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.412134 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.412181 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.412193 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.412209 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.412219 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:17Z","lastTransitionTime":"2025-10-01T09:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.514538 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.514572 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.514581 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.514594 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.514603 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:17Z","lastTransitionTime":"2025-10-01T09:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.522796 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:37:17 crc kubenswrapper[4787]: E1001 09:37:17.522997 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.531826 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.618049 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.618112 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.618123 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.618138 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.618150 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:17Z","lastTransitionTime":"2025-10-01T09:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.721512 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.721620 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.721636 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.721656 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.721673 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:17Z","lastTransitionTime":"2025-10-01T09:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.824482 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.824528 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.824540 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.824555 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.824566 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:17Z","lastTransitionTime":"2025-10-01T09:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.926642 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.926681 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.926690 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.926702 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:17 crc kubenswrapper[4787]: I1001 09:37:17.926711 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:17Z","lastTransitionTime":"2025-10-01T09:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.029174 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.029215 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.029223 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.029237 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.029247 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:18Z","lastTransitionTime":"2025-10-01T09:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.131337 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.131374 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.131383 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.131395 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.131404 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:18Z","lastTransitionTime":"2025-10-01T09:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.233778 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.233820 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.233830 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.233843 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.233853 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:18Z","lastTransitionTime":"2025-10-01T09:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.336749 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.336816 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.336829 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.336844 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.336859 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:18Z","lastTransitionTime":"2025-10-01T09:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.439784 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.439829 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.439842 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.439858 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.439871 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:18Z","lastTransitionTime":"2025-10-01T09:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.523141 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.523176 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.523212 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:37:18 crc kubenswrapper[4787]: E1001 09:37:18.523276 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:37:18 crc kubenswrapper[4787]: E1001 09:37:18.523349 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:37:18 crc kubenswrapper[4787]: E1001 09:37:18.523408 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.534130 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"180a8b2f-41ef-460c-98f2-2914e3b8f9d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f445035e4b6faf9fca8e0dbe8c69668de8825ccebd6ace48f80685c88312b707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f63e01c8f45be048cd6ade60cb51f054fbea42bb51d7b5cd32f71cf1f15ce21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1d95bbc5f966defa47192b7cc53c32f75894f4b51ce8411d7bf828e04c282b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c3cd7c1e19e00dbc9cb8cb8f3f857ae636ff3818c589769f15aa87cb94327ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c3cd7c1e19e00dbc9cb8cb8f3f857ae636ff3818c589769f15aa87cb94327ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.542265 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.542302 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.542312 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.542326 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.542339 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:18Z","lastTransitionTime":"2025-10-01T09:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.544971 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc37698-af76-4818-b216-ffe35af67929\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1bb16edb25b63e957cc0bceb04a9fa9244e68881cb45be7c7d627c5faa74232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://995030421e0cbfeb7baa8ccb574d269dd178f8185bcaa9000fe5e7fc221b1188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d4c78dd607b81a28bcfe0ffa69c02628cf06aff5fd6a574069639c5bac044a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://144123d273f3f1af4fb857d731a16d9891b097c98229a035eb132f360fa8b2f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.556823 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72701c6061105b7bb81545b5f0b035b220b22b2f8009e82809a037870c625685\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.566977 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.578434 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c830596f83deb4fa998fa9af881c7eb1f3673588f444edc9cbf895c19b8248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"message\\\":\\\"2025-10-01T09:36:30+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_54bdebc1-b80c-438d-921a-17ba60e91a86\\\\n2025-10-01T09:36:30+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_54bdebc1-b80c-438d-921a-17ba60e91a86 to /host/opt/cni/bin/\\\\n2025-10-01T09:36:30Z [verbose] multus-daemon started\\\\n2025-10-01T09:36:30Z [verbose] Readiness Indicator file check\\\\n2025-10-01T09:37:15Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.588456 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.596904 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebc8a6e2b9c6104b0f773e99d68a62e1b2bf0adc05e6f291cf91e7a8cf1bc121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.605456 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"049825a6-0cec-4adf-8b77-4889bdc4d56f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18044055eb1f6cfd7cb05f330314231b1c563c34a51ba5ed79073988eb563cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc8192c84986df76c9a45deba61aaaa1b0a019fce0000e1c9916c4fe31d1ba1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-84q9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.614588 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d95f8b1aac879d1746d8a8af7e1d5e672b36c31d5b95c23fee70b237bc3716e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.623524 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47825359-aa6e-4315-9a38-685f11f590f7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93d6842d7f8928e5a2b557a70568f5b895e17f9be42d3fadae0b91ba692f5fd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d5cf849bb15a6e2c7abb11fc455dd38be5e20a59158454f27ef69541ff50765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d5cf849bb15a6e2c7abb11fc455dd38be5e20a59158454f27ef69541ff50765\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.634479 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.644556 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.644584 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.644592 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.644605 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.644614 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:18Z","lastTransitionTime":"2025-10-01T09:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.646046 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4990da342a20f6ab1e97774294bd6856c49ca17de21d893a638c13c8e727dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.658743 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b093e96a244fdec4c5cdce62650d4da31c2ec46207a9c08bbd0a28bf286abb52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4664cb1dde25ada88989e81f0319aedc6dbd363f608a50623efaa3eeb8d00763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.672788 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0814bed7-1117-4dff-97ef-a3cd1240e615\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c91542e7254c8289b734b3ffa8e8226f2445c915c264e09c179b99604abad1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:36:27Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1001 09:36:22.478145 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 09:36:22.478910 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3375976971/tls.crt::/tmp/serving-cert-3375976971/tls.key\\\\\\\"\\\\nI1001 09:36:27.943129 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 09:36:27.955126 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 09:36:27.955153 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 09:36:27.955179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 09:36:27.955185 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 09:36:27.963955 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1001 09:36:27.963979 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1001 09:36:27.963986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964008 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 09:36:27.964015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 09:36:27.964018 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 09:36:27.964022 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1001 09:36:27.966112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.683766 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7aa5c938897ea621fc7c71c86e6a1dd2acca273e13afe71e9a27fc81f34cf2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.698143 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c68b374f72c2a689eb09480e7ae3e88691daf0cb380a52f655eb1c651a66a281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.717436 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e77339fe976956bb53dcd610364af111550cb6bfa1103daf7f706a9ee74c4439\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e77339fe976956bb53dcd610364af111550cb6bfa1103daf7f706a9ee74c4439\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:36:56Z\\\",\\\"message\\\":\\\"{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: kube-storage-version-migrator-operator,},ClusterIP:10.217.5.36,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.36],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1001 09:36:56.575330 6417 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dmsxm_openshift-ovn-kubernetes(eadf2545-886e-4642-b909-704c6cd4134e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.727346 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6zz6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb4a902-9efa-4088-a236-9a634fd085fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k5bxp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k5bxp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6zz6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:18Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.746944 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.746987 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.746999 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.747013 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.747025 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:18Z","lastTransitionTime":"2025-10-01T09:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.849397 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.849447 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.849456 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.849471 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.849481 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:18Z","lastTransitionTime":"2025-10-01T09:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.951458 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.951497 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.951507 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.951522 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:18 crc kubenswrapper[4787]: I1001 09:37:18.951532 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:18Z","lastTransitionTime":"2025-10-01T09:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.053992 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.054031 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.054042 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.054058 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.054069 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:19Z","lastTransitionTime":"2025-10-01T09:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.156262 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.156327 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.156342 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.156363 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.156378 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:19Z","lastTransitionTime":"2025-10-01T09:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.258662 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.258731 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.258754 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.258787 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.258808 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:19Z","lastTransitionTime":"2025-10-01T09:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.361578 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.361662 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.361685 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.361717 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.361739 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:19Z","lastTransitionTime":"2025-10-01T09:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.465500 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.465550 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.465560 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.465575 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.465584 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:19Z","lastTransitionTime":"2025-10-01T09:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.522827 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:37:19 crc kubenswrapper[4787]: E1001 09:37:19.523019 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.568240 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.568305 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.568314 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.568328 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.568338 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:19Z","lastTransitionTime":"2025-10-01T09:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.670877 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.670922 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.670931 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.670944 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.670956 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:19Z","lastTransitionTime":"2025-10-01T09:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.772651 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.772719 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.772739 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.772764 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.772781 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:19Z","lastTransitionTime":"2025-10-01T09:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.875600 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.875686 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.875699 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.875717 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.875729 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:19Z","lastTransitionTime":"2025-10-01T09:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.977925 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.977970 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.977980 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.978000 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:19 crc kubenswrapper[4787]: I1001 09:37:19.978012 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:19Z","lastTransitionTime":"2025-10-01T09:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.079975 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.080018 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.080035 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.080050 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.080062 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:20Z","lastTransitionTime":"2025-10-01T09:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.182947 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.182995 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.183006 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.183022 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.183032 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:20Z","lastTransitionTime":"2025-10-01T09:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.284820 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.284855 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.284865 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.284879 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.284888 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:20Z","lastTransitionTime":"2025-10-01T09:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.387412 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.387451 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.387459 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.387477 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.387485 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:20Z","lastTransitionTime":"2025-10-01T09:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.489699 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.489742 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.489753 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.489767 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.489778 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:20Z","lastTransitionTime":"2025-10-01T09:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.523424 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.523510 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:37:20 crc kubenswrapper[4787]: E1001 09:37:20.523553 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.523424 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:37:20 crc kubenswrapper[4787]: E1001 09:37:20.523654 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:37:20 crc kubenswrapper[4787]: E1001 09:37:20.523746 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.524323 4787 scope.go:117] "RemoveContainer" containerID="e77339fe976956bb53dcd610364af111550cb6bfa1103daf7f706a9ee74c4439" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.591399 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.591432 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.591441 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.591473 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.591483 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:20Z","lastTransitionTime":"2025-10-01T09:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.693526 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.693558 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.693579 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.693595 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.693603 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:20Z","lastTransitionTime":"2025-10-01T09:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.795728 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.795767 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.795776 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.795788 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.795796 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:20Z","lastTransitionTime":"2025-10-01T09:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.898352 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.898389 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.898403 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.898426 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.898437 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:20Z","lastTransitionTime":"2025-10-01T09:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.928615 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dmsxm_eadf2545-886e-4642-b909-704c6cd4134e/ovnkube-controller/2.log" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.930799 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" event={"ID":"eadf2545-886e-4642-b909-704c6cd4134e","Type":"ContainerStarted","Data":"c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122"} Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.931383 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.954062 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72701c6061105b7bb81545b5f0b035b220b22b2f8009e82809a037870c625685\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:20Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.970067 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:20Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:20 crc kubenswrapper[4787]: I1001 09:37:20.999409 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c830596f83deb4fa998fa9af881c7eb1f3673588f444edc9cbf895c19b8248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"message\\\":\\\"2025-10-01T09:36:30+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_54bdebc1-b80c-438d-921a-17ba60e91a86\\\\n2025-10-01T09:36:30+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_54bdebc1-b80c-438d-921a-17ba60e91a86 to /host/opt/cni/bin/\\\\n2025-10-01T09:36:30Z [verbose] multus-daemon started\\\\n2025-10-01T09:36:30Z [verbose] Readiness Indicator file check\\\\n2025-10-01T09:37:15Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:20Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.000462 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.000525 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.000539 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.000557 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.000570 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:21Z","lastTransitionTime":"2025-10-01T09:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.009644 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"180a8b2f-41ef-460c-98f2-2914e3b8f9d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f445035e4b6faf9fca8e0dbe8c69668de8825ccebd6ace48f80685c88312b707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f63e01c8f45be048cd6ade60cb51f054fbea42bb51d7b5cd32f71cf1f15ce21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1d95bbc5f966defa47192b7cc53c32f75894f4b51ce8411d7bf828e04c282b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c3cd7c1e19e00dbc9cb8cb8f3f857ae636ff3818c589769f15aa87cb94327ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c3cd7c1e19e00dbc9cb8cb8f3f857ae636ff3818c589769f15aa87cb94327ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.020876 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc37698-af76-4818-b216-ffe35af67929\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1bb16edb25b63e957cc0bceb04a9fa9244e68881cb45be7c7d627c5faa74232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://995030421e0cbfeb7baa8ccb574d269dd178f8185bcaa9000fe5e7fc221b1188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d4c78dd607b81a28bcfe0ffa69c02628cf06aff5fd6a574069639c5bac044a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://144123d273f3f1af4fb857d731a16d9891b097c98229a035eb132f360fa8b2f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.031496 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"049825a6-0cec-4adf-8b77-4889bdc4d56f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18044055eb1f6cfd7cb05f330314231b1c563c34a51ba5ed79073988eb563cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc8192c84986df76c9a45deba61aaaa1b0a019fce0000e1c9916c4fe31d1ba1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-84q9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.041561 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.055155 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebc8a6e2b9c6104b0f773e99d68a62e1b2bf0adc05e6f291cf91e7a8cf1bc121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.066481 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.077350 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4990da342a20f6ab1e97774294bd6856c49ca17de21d893a638c13c8e727dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.088524 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b093e96a244fdec4c5cdce62650d4da31c2ec46207a9c08bbd0a28bf286abb52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4664cb1dde25ada88989e81f0319aedc6dbd363f608a50623efaa3eeb8d00763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.097412 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d95f8b1aac879d1746d8a8af7e1d5e672b36c31d5b95c23fee70b237bc3716e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.102926 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.102995 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.103004 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.103017 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.103027 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:21Z","lastTransitionTime":"2025-10-01T09:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.105401 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47825359-aa6e-4315-9a38-685f11f590f7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93d6842d7f8928e5a2b557a70568f5b895e17f9be42d3fadae0b91ba692f5fd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d5cf849bb15a6e2c7abb11fc455dd38be5e20a59158454f27ef69541ff50765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d5cf849bb15a6e2c7abb11fc455dd38be5e20a59158454f27ef69541ff50765\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.117152 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c68b374f72c2a689eb09480e7ae3e88691daf0cb380a52f655eb1c651a66a281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.135663 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e77339fe976956bb53dcd610364af111550cb6bfa1103daf7f706a9ee74c4439\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:36:56Z\\\",\\\"message\\\":\\\"{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: kube-storage-version-migrator-operator,},ClusterIP:10.217.5.36,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.36],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1001 09:36:56.575330 6417 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.147641 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6zz6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb4a902-9efa-4088-a236-9a634fd085fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k5bxp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k5bxp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6zz6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.161905 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0814bed7-1117-4dff-97ef-a3cd1240e615\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c91542e7254c8289b734b3ffa8e8226f2445c915c264e09c179b99604abad1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:36:27Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1001 09:36:22.478145 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 09:36:22.478910 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3375976971/tls.crt::/tmp/serving-cert-3375976971/tls.key\\\\\\\"\\\\nI1001 09:36:27.943129 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 09:36:27.955126 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 09:36:27.955153 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 09:36:27.955179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 09:36:27.955185 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 09:36:27.963955 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1001 09:36:27.963979 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1001 09:36:27.963986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964008 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 09:36:27.964015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 09:36:27.964018 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 09:36:27.964022 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1001 09:36:27.966112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.171263 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7aa5c938897ea621fc7c71c86e6a1dd2acca273e13afe71e9a27fc81f34cf2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.204949 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.204988 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.204997 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.205011 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.205020 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:21Z","lastTransitionTime":"2025-10-01T09:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.307354 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.307390 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.307398 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.307413 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.307422 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:21Z","lastTransitionTime":"2025-10-01T09:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.409551 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.409596 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.409607 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.409627 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.409639 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:21Z","lastTransitionTime":"2025-10-01T09:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.512000 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.512040 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.512050 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.512063 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.512089 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:21Z","lastTransitionTime":"2025-10-01T09:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.523493 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:37:21 crc kubenswrapper[4787]: E1001 09:37:21.523648 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.614447 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.614500 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.614509 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.614521 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.614531 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:21Z","lastTransitionTime":"2025-10-01T09:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.716947 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.716985 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.716996 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.717010 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.717019 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:21Z","lastTransitionTime":"2025-10-01T09:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.819980 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.820013 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.820023 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.820037 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.820051 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:21Z","lastTransitionTime":"2025-10-01T09:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.922859 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.922905 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.922916 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.922930 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.922941 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:21Z","lastTransitionTime":"2025-10-01T09:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.936747 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dmsxm_eadf2545-886e-4642-b909-704c6cd4134e/ovnkube-controller/3.log" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.937347 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dmsxm_eadf2545-886e-4642-b909-704c6cd4134e/ovnkube-controller/2.log" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.940130 4787 generic.go:334] "Generic (PLEG): container finished" podID="eadf2545-886e-4642-b909-704c6cd4134e" containerID="c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122" exitCode=1 Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.940191 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" event={"ID":"eadf2545-886e-4642-b909-704c6cd4134e","Type":"ContainerDied","Data":"c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122"} Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.940237 4787 scope.go:117] "RemoveContainer" containerID="e77339fe976956bb53dcd610364af111550cb6bfa1103daf7f706a9ee74c4439" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.940892 4787 scope.go:117] "RemoveContainer" containerID="c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122" Oct 01 09:37:21 crc kubenswrapper[4787]: E1001 09:37:21.941096 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dmsxm_openshift-ovn-kubernetes(eadf2545-886e-4642-b909-704c6cd4134e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" podUID="eadf2545-886e-4642-b909-704c6cd4134e" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.956720 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47825359-aa6e-4315-9a38-685f11f590f7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93d6842d7f8928e5a2b557a70568f5b895e17f9be42d3fadae0b91ba692f5fd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d5cf849bb15a6e2c7abb11fc455dd38be5e20a59158454f27ef69541ff50765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d5cf849bb15a6e2c7abb11fc455dd38be5e20a59158454f27ef69541ff50765\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.971925 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:21 crc kubenswrapper[4787]: I1001 09:37:21.986727 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4990da342a20f6ab1e97774294bd6856c49ca17de21d893a638c13c8e727dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.003167 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b093e96a244fdec4c5cdce62650d4da31c2ec46207a9c08bbd0a28bf286abb52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4664cb1dde25ada88989e81f0319aedc6dbd363f608a50623efaa3eeb8d00763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:21Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.014886 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d95f8b1aac879d1746d8a8af7e1d5e672b36c31d5b95c23fee70b237bc3716e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.025603 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.025633 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.025644 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.025659 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.025671 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:22Z","lastTransitionTime":"2025-10-01T09:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.028619 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0814bed7-1117-4dff-97ef-a3cd1240e615\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c91542e7254c8289b734b3ffa8e8226f2445c915c264e09c179b99604abad1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:36:27Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1001 09:36:22.478145 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 09:36:22.478910 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3375976971/tls.crt::/tmp/serving-cert-3375976971/tls.key\\\\\\\"\\\\nI1001 09:36:27.943129 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 09:36:27.955126 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 09:36:27.955153 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 09:36:27.955179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 09:36:27.955185 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 09:36:27.963955 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1001 09:36:27.963979 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1001 09:36:27.963986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964008 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 09:36:27.964015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 09:36:27.964018 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 09:36:27.964022 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1001 09:36:27.966112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.039259 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7aa5c938897ea621fc7c71c86e6a1dd2acca273e13afe71e9a27fc81f34cf2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.056070 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c68b374f72c2a689eb09480e7ae3e88691daf0cb380a52f655eb1c651a66a281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.076773 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e77339fe976956bb53dcd610364af111550cb6bfa1103daf7f706a9ee74c4439\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:36:56Z\\\",\\\"message\\\":\\\"{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: kube-storage-version-migrator-operator,},ClusterIP:10.217.5.36,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.36],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1001 09:36:56.575330 6417 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:37:21Z\\\",\\\"message\\\":\\\"ift-image-registry/node-ca-5ghjd\\\\nF1001 09:37:21.403568 6768 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:21Z is after 2025-08-24T17:21:41Z]\\\\nI1001 09:37:21.403587 6768 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m\\\\nI1001 09:37:21.403590 6768 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-5ghjd\\\\nI1001 09:37:21.403594 6768 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1001 09:37:21.403595 6768 ovn.go:134] Ensuring zone local for Pod openshift-ovn-k\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:37:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.090552 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6zz6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb4a902-9efa-4088-a236-9a634fd085fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k5bxp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k5bxp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6zz6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.102794 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"180a8b2f-41ef-460c-98f2-2914e3b8f9d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f445035e4b6faf9fca8e0dbe8c69668de8825ccebd6ace48f80685c88312b707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f63e01c8f45be048cd6ade60cb51f054fbea42bb51d7b5cd32f71cf1f15ce21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1d95bbc5f966defa47192b7cc53c32f75894f4b51ce8411d7bf828e04c282b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c3cd7c1e19e00dbc9cb8cb8f3f857ae636ff3818c589769f15aa87cb94327ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c3cd7c1e19e00dbc9cb8cb8f3f857ae636ff3818c589769f15aa87cb94327ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.119691 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc37698-af76-4818-b216-ffe35af67929\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1bb16edb25b63e957cc0bceb04a9fa9244e68881cb45be7c7d627c5faa74232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://995030421e0cbfeb7baa8ccb574d269dd178f8185bcaa9000fe5e7fc221b1188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d4c78dd607b81a28bcfe0ffa69c02628cf06aff5fd6a574069639c5bac044a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://144123d273f3f1af4fb857d731a16d9891b097c98229a035eb132f360fa8b2f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.137421 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.137480 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.137493 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.137518 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.137532 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:22Z","lastTransitionTime":"2025-10-01T09:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.140833 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72701c6061105b7bb81545b5f0b035b220b22b2f8009e82809a037870c625685\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.157702 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.174232 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c830596f83deb4fa998fa9af881c7eb1f3673588f444edc9cbf895c19b8248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"message\\\":\\\"2025-10-01T09:36:30+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_54bdebc1-b80c-438d-921a-17ba60e91a86\\\\n2025-10-01T09:36:30+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_54bdebc1-b80c-438d-921a-17ba60e91a86 to /host/opt/cni/bin/\\\\n2025-10-01T09:36:30Z [verbose] multus-daemon started\\\\n2025-10-01T09:36:30Z [verbose] Readiness Indicator file check\\\\n2025-10-01T09:37:15Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.190724 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.200904 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebc8a6e2b9c6104b0f773e99d68a62e1b2bf0adc05e6f291cf91e7a8cf1bc121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.211514 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"049825a6-0cec-4adf-8b77-4889bdc4d56f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18044055eb1f6cfd7cb05f330314231b1c563c34a51ba5ed79073988eb563cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc8192c84986df76c9a45deba61aaaa1b0a019fce0000e1c9916c4fe31d1ba1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-84q9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.239941 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.239988 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.240000 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.240020 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.240032 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:22Z","lastTransitionTime":"2025-10-01T09:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.343384 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.343435 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.343446 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.343468 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.343483 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:22Z","lastTransitionTime":"2025-10-01T09:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.446943 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.447013 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.447038 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.447113 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.447173 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:22Z","lastTransitionTime":"2025-10-01T09:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.523058 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.523061 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:37:22 crc kubenswrapper[4787]: E1001 09:37:22.523222 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.523379 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:37:22 crc kubenswrapper[4787]: E1001 09:37:22.523455 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:37:22 crc kubenswrapper[4787]: E1001 09:37:22.523550 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.549870 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.549934 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.549950 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.549970 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.549984 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:22Z","lastTransitionTime":"2025-10-01T09:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.652823 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.652857 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.652868 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.652882 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.652894 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:22Z","lastTransitionTime":"2025-10-01T09:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.756558 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.756598 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.756611 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.756626 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.756638 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:22Z","lastTransitionTime":"2025-10-01T09:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.859096 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.859172 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.859184 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.859199 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.859209 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:22Z","lastTransitionTime":"2025-10-01T09:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.948188 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dmsxm_eadf2545-886e-4642-b909-704c6cd4134e/ovnkube-controller/3.log" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.954988 4787 scope.go:117] "RemoveContainer" containerID="c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122" Oct 01 09:37:22 crc kubenswrapper[4787]: E1001 09:37:22.955500 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dmsxm_openshift-ovn-kubernetes(eadf2545-886e-4642-b909-704c6cd4134e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" podUID="eadf2545-886e-4642-b909-704c6cd4134e" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.962568 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.962625 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.962640 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.962663 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.962679 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:22Z","lastTransitionTime":"2025-10-01T09:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.971452 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc37698-af76-4818-b216-ffe35af67929\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1bb16edb25b63e957cc0bceb04a9fa9244e68881cb45be7c7d627c5faa74232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://995030421e0cbfeb7baa8ccb574d269dd178f8185bcaa9000fe5e7fc221b1188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d4c78dd607b81a28bcfe0ffa69c02628cf06aff5fd6a574069639c5bac044a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://144123d273f3f1af4fb857d731a16d9891b097c98229a035eb132f360fa8b2f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:22 crc kubenswrapper[4787]: I1001 09:37:22.984324 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72701c6061105b7bb81545b5f0b035b220b22b2f8009e82809a037870c625685\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:22Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.002472 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.017113 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c830596f83deb4fa998fa9af881c7eb1f3673588f444edc9cbf895c19b8248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"message\\\":\\\"2025-10-01T09:36:30+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_54bdebc1-b80c-438d-921a-17ba60e91a86\\\\n2025-10-01T09:36:30+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_54bdebc1-b80c-438d-921a-17ba60e91a86 to /host/opt/cni/bin/\\\\n2025-10-01T09:36:30Z [verbose] multus-daemon started\\\\n2025-10-01T09:36:30Z [verbose] Readiness Indicator file check\\\\n2025-10-01T09:37:15Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.034881 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"180a8b2f-41ef-460c-98f2-2914e3b8f9d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f445035e4b6faf9fca8e0dbe8c69668de8825ccebd6ace48f80685c88312b707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f63e01c8f45be048cd6ade60cb51f054fbea42bb51d7b5cd32f71cf1f15ce21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1d95bbc5f966defa47192b7cc53c32f75894f4b51ce8411d7bf828e04c282b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c3cd7c1e19e00dbc9cb8cb8f3f857ae636ff3818c589769f15aa87cb94327ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c3cd7c1e19e00dbc9cb8cb8f3f857ae636ff3818c589769f15aa87cb94327ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.052003 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebc8a6e2b9c6104b0f773e99d68a62e1b2bf0adc05e6f291cf91e7a8cf1bc121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.066205 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.066252 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.066299 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.066331 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.066347 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:23Z","lastTransitionTime":"2025-10-01T09:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.070748 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"049825a6-0cec-4adf-8b77-4889bdc4d56f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18044055eb1f6cfd7cb05f330314231b1c563c34a51ba5ed79073988eb563cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc8192c84986df76c9a45deba61aaaa1b0a019fce0000e1c9916c4fe31d1ba1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-84q9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.090974 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.107527 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47825359-aa6e-4315-9a38-685f11f590f7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93d6842d7f8928e5a2b557a70568f5b895e17f9be42d3fadae0b91ba692f5fd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d5cf849bb15a6e2c7abb11fc455dd38be5e20a59158454f27ef69541ff50765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d5cf849bb15a6e2c7abb11fc455dd38be5e20a59158454f27ef69541ff50765\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.122179 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.136382 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4990da342a20f6ab1e97774294bd6856c49ca17de21d893a638c13c8e727dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.152836 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b093e96a244fdec4c5cdce62650d4da31c2ec46207a9c08bbd0a28bf286abb52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4664cb1dde25ada88989e81f0319aedc6dbd363f608a50623efaa3eeb8d00763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.169698 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.169744 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.169754 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.169772 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.169783 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:23Z","lastTransitionTime":"2025-10-01T09:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.169827 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d95f8b1aac879d1746d8a8af7e1d5e672b36c31d5b95c23fee70b237bc3716e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.183009 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7aa5c938897ea621fc7c71c86e6a1dd2acca273e13afe71e9a27fc81f34cf2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.199948 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c68b374f72c2a689eb09480e7ae3e88691daf0cb380a52f655eb1c651a66a281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.219461 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:37:21Z\\\",\\\"message\\\":\\\"ift-image-registry/node-ca-5ghjd\\\\nF1001 09:37:21.403568 6768 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:21Z is after 2025-08-24T17:21:41Z]\\\\nI1001 09:37:21.403587 6768 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m\\\\nI1001 09:37:21.403590 6768 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-5ghjd\\\\nI1001 09:37:21.403594 6768 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1001 09:37:21.403595 6768 ovn.go:134] Ensuring zone local for Pod openshift-ovn-k\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:37:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dmsxm_openshift-ovn-kubernetes(eadf2545-886e-4642-b909-704c6cd4134e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.233211 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6zz6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb4a902-9efa-4088-a236-9a634fd085fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k5bxp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k5bxp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6zz6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.248918 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0814bed7-1117-4dff-97ef-a3cd1240e615\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c91542e7254c8289b734b3ffa8e8226f2445c915c264e09c179b99604abad1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:36:27Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1001 09:36:22.478145 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 09:36:22.478910 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3375976971/tls.crt::/tmp/serving-cert-3375976971/tls.key\\\\\\\"\\\\nI1001 09:36:27.943129 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 09:36:27.955126 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 09:36:27.955153 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 09:36:27.955179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 09:36:27.955185 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 09:36:27.963955 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1001 09:36:27.963979 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1001 09:36:27.963986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964008 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 09:36:27.964015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 09:36:27.964018 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 09:36:27.964022 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1001 09:36:27.966112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:23Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.273375 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.273423 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.273434 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.273454 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.273464 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:23Z","lastTransitionTime":"2025-10-01T09:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.377025 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.377135 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.377151 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.377172 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.377182 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:23Z","lastTransitionTime":"2025-10-01T09:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.480124 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.480538 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.480639 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.480770 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.480871 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:23Z","lastTransitionTime":"2025-10-01T09:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.523163 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:37:23 crc kubenswrapper[4787]: E1001 09:37:23.523304 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.583869 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.583919 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.583935 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.583961 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.583973 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:23Z","lastTransitionTime":"2025-10-01T09:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.687223 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.687257 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.687265 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.687279 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.687288 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:23Z","lastTransitionTime":"2025-10-01T09:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.791118 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.791187 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.791205 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.791233 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.791253 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:23Z","lastTransitionTime":"2025-10-01T09:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.894509 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.894596 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.894616 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.894655 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.894690 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:23Z","lastTransitionTime":"2025-10-01T09:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.998216 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.998271 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.998283 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.998302 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:23 crc kubenswrapper[4787]: I1001 09:37:23.998317 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:23Z","lastTransitionTime":"2025-10-01T09:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.102715 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.102806 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.102832 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.102869 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.102892 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:24Z","lastTransitionTime":"2025-10-01T09:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.206014 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.206107 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.206127 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.206154 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.206174 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:24Z","lastTransitionTime":"2025-10-01T09:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.310472 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.310565 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.310592 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.310628 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.310657 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:24Z","lastTransitionTime":"2025-10-01T09:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.414955 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.414994 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.415003 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.415020 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.415034 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:24Z","lastTransitionTime":"2025-10-01T09:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.518005 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.518051 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.518064 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.518099 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.518112 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:24Z","lastTransitionTime":"2025-10-01T09:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.523499 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.523609 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.523884 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:37:24 crc kubenswrapper[4787]: E1001 09:37:24.523980 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:37:24 crc kubenswrapper[4787]: E1001 09:37:24.523867 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:37:24 crc kubenswrapper[4787]: E1001 09:37:24.524033 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.620714 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.620840 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.621177 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.621220 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.621246 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:24Z","lastTransitionTime":"2025-10-01T09:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.725181 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.725252 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.725283 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.725321 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.725349 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:24Z","lastTransitionTime":"2025-10-01T09:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.828774 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.828836 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.828849 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.828877 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.828892 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:24Z","lastTransitionTime":"2025-10-01T09:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.933196 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.933292 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.933317 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.933356 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:24 crc kubenswrapper[4787]: I1001 09:37:24.933382 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:24Z","lastTransitionTime":"2025-10-01T09:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.037146 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.037223 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.037250 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.037329 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.037361 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:25Z","lastTransitionTime":"2025-10-01T09:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.140726 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.140839 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.140859 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.140929 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.140952 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:25Z","lastTransitionTime":"2025-10-01T09:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.244955 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.245008 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.245024 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.245047 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.245064 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:25Z","lastTransitionTime":"2025-10-01T09:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.347550 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.347599 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.347611 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.347627 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.347638 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:25Z","lastTransitionTime":"2025-10-01T09:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.452351 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.452418 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.452438 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.452465 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.452486 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:25Z","lastTransitionTime":"2025-10-01T09:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.485796 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.485851 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.485864 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.485885 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.485900 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:25Z","lastTransitionTime":"2025-10-01T09:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:25 crc kubenswrapper[4787]: E1001 09:37:25.499846 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:25Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.505704 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.505752 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.505761 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.505779 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.505796 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:25Z","lastTransitionTime":"2025-10-01T09:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:25 crc kubenswrapper[4787]: E1001 09:37:25.522375 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:25Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.522744 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:37:25 crc kubenswrapper[4787]: E1001 09:37:25.522906 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.527859 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.527993 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.528020 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.528105 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.528133 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:25Z","lastTransitionTime":"2025-10-01T09:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:25 crc kubenswrapper[4787]: E1001 09:37:25.546152 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:25Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.551193 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.551246 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.551264 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.551287 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.551309 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:25Z","lastTransitionTime":"2025-10-01T09:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:25 crc kubenswrapper[4787]: E1001 09:37:25.565666 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:25Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.576707 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.576748 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.576796 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.576818 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.576833 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:25Z","lastTransitionTime":"2025-10-01T09:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:25 crc kubenswrapper[4787]: E1001 09:37:25.593785 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:25Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:25 crc kubenswrapper[4787]: E1001 09:37:25.594224 4787 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.596516 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.596576 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.596587 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.596603 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.596613 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:25Z","lastTransitionTime":"2025-10-01T09:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.700432 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.700527 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.700553 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.700595 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.700623 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:25Z","lastTransitionTime":"2025-10-01T09:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.803871 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.803939 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.803953 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.803974 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.803991 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:25Z","lastTransitionTime":"2025-10-01T09:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.907816 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.907894 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.907915 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.907944 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:25 crc kubenswrapper[4787]: I1001 09:37:25.907965 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:25Z","lastTransitionTime":"2025-10-01T09:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.011291 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.011385 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.011414 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.011449 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.011475 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:26Z","lastTransitionTime":"2025-10-01T09:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.115493 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.115571 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.115588 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.115622 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.115640 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:26Z","lastTransitionTime":"2025-10-01T09:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.220978 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.221058 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.221129 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.221157 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.221178 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:26Z","lastTransitionTime":"2025-10-01T09:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.324801 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.324844 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.324853 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.324869 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.324880 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:26Z","lastTransitionTime":"2025-10-01T09:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.427799 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.428591 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.428610 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.428643 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.428665 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:26Z","lastTransitionTime":"2025-10-01T09:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.523666 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.523712 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.523685 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:37:26 crc kubenswrapper[4787]: E1001 09:37:26.523912 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:37:26 crc kubenswrapper[4787]: E1001 09:37:26.524140 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:37:26 crc kubenswrapper[4787]: E1001 09:37:26.524243 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.531224 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.531287 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.531312 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.531343 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.531368 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:26Z","lastTransitionTime":"2025-10-01T09:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.635235 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.635314 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.635327 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.635351 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.635363 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:26Z","lastTransitionTime":"2025-10-01T09:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.738954 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.739012 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.739022 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.739041 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.739052 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:26Z","lastTransitionTime":"2025-10-01T09:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.842233 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.842290 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.842306 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.842324 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.842335 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:26Z","lastTransitionTime":"2025-10-01T09:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.945061 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.945131 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.945144 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.945159 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:26 crc kubenswrapper[4787]: I1001 09:37:26.945170 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:26Z","lastTransitionTime":"2025-10-01T09:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.047797 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.047875 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.047892 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.047914 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.047929 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:27Z","lastTransitionTime":"2025-10-01T09:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.151465 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.151529 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.151542 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.151608 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.151624 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:27Z","lastTransitionTime":"2025-10-01T09:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.255226 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.255302 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.255321 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.255350 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.255372 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:27Z","lastTransitionTime":"2025-10-01T09:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.358886 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.358967 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.359004 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.359025 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.359037 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:27Z","lastTransitionTime":"2025-10-01T09:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.462414 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.462482 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.462492 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.462510 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.462521 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:27Z","lastTransitionTime":"2025-10-01T09:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.523357 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:37:27 crc kubenswrapper[4787]: E1001 09:37:27.523616 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.565856 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.565902 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.565912 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.565928 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.565937 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:27Z","lastTransitionTime":"2025-10-01T09:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.668404 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.668452 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.668473 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.668501 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.668526 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:27Z","lastTransitionTime":"2025-10-01T09:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.771208 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.771270 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.771286 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.771310 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.771327 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:27Z","lastTransitionTime":"2025-10-01T09:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.875038 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.875616 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.875630 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.875652 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.875665 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:27Z","lastTransitionTime":"2025-10-01T09:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.978646 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.978759 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.978784 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.978849 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:27 crc kubenswrapper[4787]: I1001 09:37:27.978873 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:27Z","lastTransitionTime":"2025-10-01T09:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.082278 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.082360 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.082389 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.082421 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.082441 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:28Z","lastTransitionTime":"2025-10-01T09:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.188370 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.188435 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.188449 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.188469 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.188489 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:28Z","lastTransitionTime":"2025-10-01T09:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.291010 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.291044 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.291053 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.291065 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.291093 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:28Z","lastTransitionTime":"2025-10-01T09:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.393355 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.393393 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.393404 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.393421 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.393432 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:28Z","lastTransitionTime":"2025-10-01T09:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.497257 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.497319 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.497338 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.497363 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.497381 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:28Z","lastTransitionTime":"2025-10-01T09:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.523442 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.523597 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.523692 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:37:28 crc kubenswrapper[4787]: E1001 09:37:28.523629 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:37:28 crc kubenswrapper[4787]: E1001 09:37:28.523925 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:37:28 crc kubenswrapper[4787]: E1001 09:37:28.524239 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.549201 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0814bed7-1117-4dff-97ef-a3cd1240e615\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c91542e7254c8289b734b3ffa8e8226f2445c915c264e09c179b99604abad1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:36:27Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1001 09:36:22.478145 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 09:36:22.478910 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3375976971/tls.crt::/tmp/serving-cert-3375976971/tls.key\\\\\\\"\\\\nI1001 09:36:27.943129 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 09:36:27.955126 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 09:36:27.955153 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 09:36:27.955179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 09:36:27.955185 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 09:36:27.963955 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1001 09:36:27.963979 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1001 09:36:27.963986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964008 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 09:36:27.964015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 09:36:27.964018 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 09:36:27.964022 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1001 09:36:27.966112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:28Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.563426 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7aa5c938897ea621fc7c71c86e6a1dd2acca273e13afe71e9a27fc81f34cf2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:28Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.588749 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c68b374f72c2a689eb09480e7ae3e88691daf0cb380a52f655eb1c651a66a281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:28Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.601013 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.601119 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.601170 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.601192 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.601204 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:28Z","lastTransitionTime":"2025-10-01T09:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.611733 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:37:21Z\\\",\\\"message\\\":\\\"ift-image-registry/node-ca-5ghjd\\\\nF1001 09:37:21.403568 6768 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:21Z is after 2025-08-24T17:21:41Z]\\\\nI1001 09:37:21.403587 6768 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m\\\\nI1001 09:37:21.403590 6768 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-5ghjd\\\\nI1001 09:37:21.403594 6768 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1001 09:37:21.403595 6768 ovn.go:134] Ensuring zone local for Pod openshift-ovn-k\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:37:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dmsxm_openshift-ovn-kubernetes(eadf2545-886e-4642-b909-704c6cd4134e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:28Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.628028 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6zz6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb4a902-9efa-4088-a236-9a634fd085fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k5bxp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k5bxp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6zz6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:28Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.642408 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"180a8b2f-41ef-460c-98f2-2914e3b8f9d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f445035e4b6faf9fca8e0dbe8c69668de8825ccebd6ace48f80685c88312b707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f63e01c8f45be048cd6ade60cb51f054fbea42bb51d7b5cd32f71cf1f15ce21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1d95bbc5f966defa47192b7cc53c32f75894f4b51ce8411d7bf828e04c282b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c3cd7c1e19e00dbc9cb8cb8f3f857ae636ff3818c589769f15aa87cb94327ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c3cd7c1e19e00dbc9cb8cb8f3f857ae636ff3818c589769f15aa87cb94327ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:28Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.656666 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc37698-af76-4818-b216-ffe35af67929\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1bb16edb25b63e957cc0bceb04a9fa9244e68881cb45be7c7d627c5faa74232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://995030421e0cbfeb7baa8ccb574d269dd178f8185bcaa9000fe5e7fc221b1188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d4c78dd607b81a28bcfe0ffa69c02628cf06aff5fd6a574069639c5bac044a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://144123d273f3f1af4fb857d731a16d9891b097c98229a035eb132f360fa8b2f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:28Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.673595 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72701c6061105b7bb81545b5f0b035b220b22b2f8009e82809a037870c625685\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:28Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.694018 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:28Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.705293 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.705348 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.705362 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.705382 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.705398 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:28Z","lastTransitionTime":"2025-10-01T09:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.711264 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c830596f83deb4fa998fa9af881c7eb1f3673588f444edc9cbf895c19b8248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"message\\\":\\\"2025-10-01T09:36:30+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_54bdebc1-b80c-438d-921a-17ba60e91a86\\\\n2025-10-01T09:36:30+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_54bdebc1-b80c-438d-921a-17ba60e91a86 to /host/opt/cni/bin/\\\\n2025-10-01T09:36:30Z [verbose] multus-daemon started\\\\n2025-10-01T09:36:30Z [verbose] Readiness Indicator file check\\\\n2025-10-01T09:37:15Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:28Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.725576 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:28Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.742385 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebc8a6e2b9c6104b0f773e99d68a62e1b2bf0adc05e6f291cf91e7a8cf1bc121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:28Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.761127 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"049825a6-0cec-4adf-8b77-4889bdc4d56f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18044055eb1f6cfd7cb05f330314231b1c563c34a51ba5ed79073988eb563cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc8192c84986df76c9a45deba61aaaa1b0a019fce0000e1c9916c4fe31d1ba1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-84q9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:28Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.775972 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d95f8b1aac879d1746d8a8af7e1d5e672b36c31d5b95c23fee70b237bc3716e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:28Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.794130 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47825359-aa6e-4315-9a38-685f11f590f7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93d6842d7f8928e5a2b557a70568f5b895e17f9be42d3fadae0b91ba692f5fd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d5cf849bb15a6e2c7abb11fc455dd38be5e20a59158454f27ef69541ff50765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d5cf849bb15a6e2c7abb11fc455dd38be5e20a59158454f27ef69541ff50765\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:28Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.808765 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.808819 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.808833 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.808851 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.808863 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:28Z","lastTransitionTime":"2025-10-01T09:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.813449 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:28Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.827629 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4990da342a20f6ab1e97774294bd6856c49ca17de21d893a638c13c8e727dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:28Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.845838 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b093e96a244fdec4c5cdce62650d4da31c2ec46207a9c08bbd0a28bf286abb52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4664cb1dde25ada88989e81f0319aedc6dbd363f608a50623efaa3eeb8d00763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:28Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.911603 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.911673 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.911693 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.911724 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:28 crc kubenswrapper[4787]: I1001 09:37:28.911745 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:28Z","lastTransitionTime":"2025-10-01T09:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.015493 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.016113 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.016134 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.016154 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.016179 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:29Z","lastTransitionTime":"2025-10-01T09:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.119787 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.119853 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.119861 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.119879 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.119889 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:29Z","lastTransitionTime":"2025-10-01T09:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.223745 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.223807 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.223819 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.223840 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.223852 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:29Z","lastTransitionTime":"2025-10-01T09:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.326419 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.326459 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.326472 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.326489 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.326500 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:29Z","lastTransitionTime":"2025-10-01T09:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.428780 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.429002 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.429065 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.429493 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.429660 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:29Z","lastTransitionTime":"2025-10-01T09:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.522988 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:37:29 crc kubenswrapper[4787]: E1001 09:37:29.523166 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.531268 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.531460 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.531525 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.531597 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.531667 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:29Z","lastTransitionTime":"2025-10-01T09:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.634462 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.634524 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.634542 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.634567 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.634586 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:29Z","lastTransitionTime":"2025-10-01T09:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.736554 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.736600 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.736612 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.736628 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.736644 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:29Z","lastTransitionTime":"2025-10-01T09:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.839629 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.839685 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.839703 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.839728 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.839792 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:29Z","lastTransitionTime":"2025-10-01T09:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.942110 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.942178 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.942193 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.942220 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:29 crc kubenswrapper[4787]: I1001 09:37:29.942235 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:29Z","lastTransitionTime":"2025-10-01T09:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.045269 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.045307 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.045315 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.045329 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.045340 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:30Z","lastTransitionTime":"2025-10-01T09:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.148160 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.148250 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.148269 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.148300 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.148320 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:30Z","lastTransitionTime":"2025-10-01T09:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.252129 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.252170 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.252181 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.252197 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.252211 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:30Z","lastTransitionTime":"2025-10-01T09:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.354914 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.354971 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.354986 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.355005 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.355019 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:30Z","lastTransitionTime":"2025-10-01T09:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.458439 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.458512 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.458534 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.458562 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.458583 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:30Z","lastTransitionTime":"2025-10-01T09:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.523784 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.523945 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.523956 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:37:30 crc kubenswrapper[4787]: E1001 09:37:30.524138 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:37:30 crc kubenswrapper[4787]: E1001 09:37:30.524335 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:37:30 crc kubenswrapper[4787]: E1001 09:37:30.524383 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.561828 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.561880 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.561893 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.561914 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.561927 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:30Z","lastTransitionTime":"2025-10-01T09:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.665804 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.665860 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.665874 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.665893 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.665905 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:30Z","lastTransitionTime":"2025-10-01T09:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.769131 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.769181 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.769192 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.769211 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.769221 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:30Z","lastTransitionTime":"2025-10-01T09:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.872373 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.872450 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.872477 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.872507 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.872528 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:30Z","lastTransitionTime":"2025-10-01T09:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.975816 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.975869 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.975880 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.975900 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:30 crc kubenswrapper[4787]: I1001 09:37:30.975911 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:30Z","lastTransitionTime":"2025-10-01T09:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.078710 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.079382 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.079478 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.079587 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.079673 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:31Z","lastTransitionTime":"2025-10-01T09:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.183327 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.183459 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.183487 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.183535 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.183560 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:31Z","lastTransitionTime":"2025-10-01T09:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.286646 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.286693 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.286704 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.286724 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.286736 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:31Z","lastTransitionTime":"2025-10-01T09:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.389201 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.389243 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.389253 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.389270 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.389283 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:31Z","lastTransitionTime":"2025-10-01T09:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.492159 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.492235 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.492249 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.492270 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.492281 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:31Z","lastTransitionTime":"2025-10-01T09:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.522703 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:37:31 crc kubenswrapper[4787]: E1001 09:37:31.522876 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.596186 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.596270 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.596295 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.596325 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.596346 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:31Z","lastTransitionTime":"2025-10-01T09:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.701148 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.701266 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.701325 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.701359 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.701419 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:31Z","lastTransitionTime":"2025-10-01T09:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.803773 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.803807 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.803877 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.803913 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.803926 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:31Z","lastTransitionTime":"2025-10-01T09:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.907817 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.907873 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.907882 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.907899 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:31 crc kubenswrapper[4787]: I1001 09:37:31.907911 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:31Z","lastTransitionTime":"2025-10-01T09:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.011241 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.011320 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.011343 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.011367 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.011384 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:32Z","lastTransitionTime":"2025-10-01T09:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.114407 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.114457 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.114468 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.114484 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.114495 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:32Z","lastTransitionTime":"2025-10-01T09:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.217659 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.218178 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.218189 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.218204 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.218218 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:32Z","lastTransitionTime":"2025-10-01T09:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.320967 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.321047 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.321070 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.321164 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.321186 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:32Z","lastTransitionTime":"2025-10-01T09:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.424961 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.425039 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.425061 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.425157 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.425194 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:32Z","lastTransitionTime":"2025-10-01T09:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.442166 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.442316 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.442359 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.442397 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.442511 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:37:32 crc kubenswrapper[4787]: E1001 09:37:32.442724 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 09:37:32 crc kubenswrapper[4787]: E1001 09:37:32.442757 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 09:37:32 crc kubenswrapper[4787]: E1001 09:37:32.442773 4787 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:37:32 crc kubenswrapper[4787]: E1001 09:37:32.442839 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-01 09:38:36.442819671 +0000 UTC m=+148.557963838 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:37:32 crc kubenswrapper[4787]: E1001 09:37:32.443130 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:36.443117518 +0000 UTC m=+148.558261685 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:37:32 crc kubenswrapper[4787]: E1001 09:37:32.443182 4787 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 09:37:32 crc kubenswrapper[4787]: E1001 09:37:32.443213 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 09:38:36.443205201 +0000 UTC m=+148.558349378 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 01 09:37:32 crc kubenswrapper[4787]: E1001 09:37:32.443413 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 01 09:37:32 crc kubenswrapper[4787]: E1001 09:37:32.443443 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 01 09:37:32 crc kubenswrapper[4787]: E1001 09:37:32.443455 4787 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:37:32 crc kubenswrapper[4787]: E1001 09:37:32.443489 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-01 09:38:36.443478868 +0000 UTC m=+148.558623035 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 01 09:37:32 crc kubenswrapper[4787]: E1001 09:37:32.443683 4787 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 09:37:32 crc kubenswrapper[4787]: E1001 09:37:32.443726 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-01 09:38:36.443716574 +0000 UTC m=+148.558860741 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.523120 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.523157 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.523233 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:37:32 crc kubenswrapper[4787]: E1001 09:37:32.523346 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:37:32 crc kubenswrapper[4787]: E1001 09:37:32.523474 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:37:32 crc kubenswrapper[4787]: E1001 09:37:32.523517 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.528519 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.528550 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.528561 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.528576 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.528586 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:32Z","lastTransitionTime":"2025-10-01T09:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.631144 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.631190 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.631201 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.631214 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.631224 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:32Z","lastTransitionTime":"2025-10-01T09:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.733515 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.733547 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.733559 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.733573 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.733584 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:32Z","lastTransitionTime":"2025-10-01T09:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.836396 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.836432 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.836441 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.836460 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.836469 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:32Z","lastTransitionTime":"2025-10-01T09:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.938969 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.939138 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.939160 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.939185 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:32 crc kubenswrapper[4787]: I1001 09:37:32.939204 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:32Z","lastTransitionTime":"2025-10-01T09:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.042599 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.042647 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.042658 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.042678 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.042693 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:33Z","lastTransitionTime":"2025-10-01T09:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.146434 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.146494 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.146512 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.146542 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.146563 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:33Z","lastTransitionTime":"2025-10-01T09:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.250789 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.250878 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.250905 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.250948 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.250970 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:33Z","lastTransitionTime":"2025-10-01T09:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.354879 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.354950 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.354967 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.355010 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.355030 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:33Z","lastTransitionTime":"2025-10-01T09:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.458442 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.458517 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.458532 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.458557 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.458576 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:33Z","lastTransitionTime":"2025-10-01T09:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.523498 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:37:33 crc kubenswrapper[4787]: E1001 09:37:33.523780 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.562400 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.562498 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.562522 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.562583 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.562605 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:33Z","lastTransitionTime":"2025-10-01T09:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.666779 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.666890 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.666910 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.666938 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.666958 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:33Z","lastTransitionTime":"2025-10-01T09:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.770361 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.770444 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.770470 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.770508 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.770536 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:33Z","lastTransitionTime":"2025-10-01T09:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.874705 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.874784 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.874803 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.874832 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.874856 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:33Z","lastTransitionTime":"2025-10-01T09:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.978558 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.978621 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.978637 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.978666 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:33 crc kubenswrapper[4787]: I1001 09:37:33.978684 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:33Z","lastTransitionTime":"2025-10-01T09:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.081990 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.082045 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.082063 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.082117 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.082138 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:34Z","lastTransitionTime":"2025-10-01T09:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.186602 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.186715 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.186738 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.186767 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.186788 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:34Z","lastTransitionTime":"2025-10-01T09:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.290630 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.290735 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.290765 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.290805 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.290831 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:34Z","lastTransitionTime":"2025-10-01T09:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.394433 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.394527 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.394561 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.394598 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.394623 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:34Z","lastTransitionTime":"2025-10-01T09:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.498726 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.498797 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.498817 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.498848 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.498870 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:34Z","lastTransitionTime":"2025-10-01T09:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.523141 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.523276 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.523169 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:37:34 crc kubenswrapper[4787]: E1001 09:37:34.523384 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:37:34 crc kubenswrapper[4787]: E1001 09:37:34.523516 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:37:34 crc kubenswrapper[4787]: E1001 09:37:34.523647 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.602974 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.603041 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.603058 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.603122 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.603149 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:34Z","lastTransitionTime":"2025-10-01T09:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.707237 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.707291 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.707305 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.707327 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.707340 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:34Z","lastTransitionTime":"2025-10-01T09:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.811378 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.811427 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.811438 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.811457 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.811470 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:34Z","lastTransitionTime":"2025-10-01T09:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.914267 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.914358 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.914382 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.914417 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:34 crc kubenswrapper[4787]: I1001 09:37:34.914438 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:34Z","lastTransitionTime":"2025-10-01T09:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.016913 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.017096 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.017115 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.017138 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.017157 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:35Z","lastTransitionTime":"2025-10-01T09:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.121007 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.121112 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.121130 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.121150 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.121162 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:35Z","lastTransitionTime":"2025-10-01T09:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.225813 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.225924 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.225937 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.225960 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.225973 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:35Z","lastTransitionTime":"2025-10-01T09:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.329120 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.329166 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.329176 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.329190 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.329200 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:35Z","lastTransitionTime":"2025-10-01T09:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.432740 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.432812 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.432832 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.432863 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.432888 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:35Z","lastTransitionTime":"2025-10-01T09:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.523532 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:37:35 crc kubenswrapper[4787]: E1001 09:37:35.523690 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.524695 4787 scope.go:117] "RemoveContainer" containerID="c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122" Oct 01 09:37:35 crc kubenswrapper[4787]: E1001 09:37:35.525058 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dmsxm_openshift-ovn-kubernetes(eadf2545-886e-4642-b909-704c6cd4134e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" podUID="eadf2545-886e-4642-b909-704c6cd4134e" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.535696 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.535781 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.535801 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.535851 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.535888 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:35Z","lastTransitionTime":"2025-10-01T09:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.640523 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.640614 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.640637 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.640671 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.640693 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:35Z","lastTransitionTime":"2025-10-01T09:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.745513 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.745606 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.745637 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.745673 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.745698 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:35Z","lastTransitionTime":"2025-10-01T09:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.780471 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.780526 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.780539 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.780559 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.780574 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:35Z","lastTransitionTime":"2025-10-01T09:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:35 crc kubenswrapper[4787]: E1001 09:37:35.794005 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:35Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.799098 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.799150 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.799160 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.799178 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.799194 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:35Z","lastTransitionTime":"2025-10-01T09:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:35 crc kubenswrapper[4787]: E1001 09:37:35.819429 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:35Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.823602 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.823631 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.823641 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.823657 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.823670 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:35Z","lastTransitionTime":"2025-10-01T09:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:35 crc kubenswrapper[4787]: E1001 09:37:35.845053 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:35Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.851286 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.851343 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.851353 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.851374 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.851390 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:35Z","lastTransitionTime":"2025-10-01T09:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:35 crc kubenswrapper[4787]: E1001 09:37:35.867573 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:35Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.872174 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.872217 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.872226 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.872241 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.872252 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:35Z","lastTransitionTime":"2025-10-01T09:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:35 crc kubenswrapper[4787]: E1001 09:37:35.887271 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:35Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:35 crc kubenswrapper[4787]: E1001 09:37:35.887515 4787 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.889209 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.889270 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.889290 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.889318 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.889339 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:35Z","lastTransitionTime":"2025-10-01T09:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.992660 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.992719 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.992740 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.992766 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:35 crc kubenswrapper[4787]: I1001 09:37:35.992786 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:35Z","lastTransitionTime":"2025-10-01T09:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.096410 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.096478 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.096497 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.096524 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.096546 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:36Z","lastTransitionTime":"2025-10-01T09:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.199516 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.199595 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.199616 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.199648 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.199672 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:36Z","lastTransitionTime":"2025-10-01T09:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.302514 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.302569 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.302581 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.302602 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.302614 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:36Z","lastTransitionTime":"2025-10-01T09:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.406316 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.406418 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.406443 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.406479 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.406505 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:36Z","lastTransitionTime":"2025-10-01T09:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.510518 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.510579 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.510588 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.510610 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.510620 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:36Z","lastTransitionTime":"2025-10-01T09:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.523209 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.523569 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:37:36 crc kubenswrapper[4787]: E1001 09:37:36.523736 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.523930 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:37:36 crc kubenswrapper[4787]: E1001 09:37:36.524277 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:37:36 crc kubenswrapper[4787]: E1001 09:37:36.524433 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.615586 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.615666 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.615682 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.615708 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.615720 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:36Z","lastTransitionTime":"2025-10-01T09:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.718963 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.719060 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.719129 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.719171 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.719196 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:36Z","lastTransitionTime":"2025-10-01T09:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.822240 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.822315 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.822335 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.822367 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.822390 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:36Z","lastTransitionTime":"2025-10-01T09:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.926585 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.926682 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.926697 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.926716 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:36 crc kubenswrapper[4787]: I1001 09:37:36.926730 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:36Z","lastTransitionTime":"2025-10-01T09:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.030966 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.031036 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.031059 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.031126 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.031150 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:37Z","lastTransitionTime":"2025-10-01T09:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.134941 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.135020 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.135044 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.135106 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.135129 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:37Z","lastTransitionTime":"2025-10-01T09:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.238146 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.238177 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.238185 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.238198 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.238208 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:37Z","lastTransitionTime":"2025-10-01T09:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.341520 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.341569 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.341578 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.341598 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.341607 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:37Z","lastTransitionTime":"2025-10-01T09:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.445663 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.445732 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.445751 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.445779 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.445799 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:37Z","lastTransitionTime":"2025-10-01T09:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.523667 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:37:37 crc kubenswrapper[4787]: E1001 09:37:37.523887 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.549493 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.549573 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.549594 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.549629 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.549652 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:37Z","lastTransitionTime":"2025-10-01T09:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.653023 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.653122 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.653142 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.653170 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.653191 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:37Z","lastTransitionTime":"2025-10-01T09:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.757207 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.757283 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.757302 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.757334 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.757356 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:37Z","lastTransitionTime":"2025-10-01T09:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.859971 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.860011 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.860021 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.860053 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.860064 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:37Z","lastTransitionTime":"2025-10-01T09:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.962855 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.962888 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.962916 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.962930 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:37 crc kubenswrapper[4787]: I1001 09:37:37.962939 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:37Z","lastTransitionTime":"2025-10-01T09:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.066064 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.066197 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.066227 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.066253 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.066272 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:38Z","lastTransitionTime":"2025-10-01T09:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.169114 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.169193 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.169216 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.169240 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.169254 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:38Z","lastTransitionTime":"2025-10-01T09:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.271325 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.271394 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.271411 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.271427 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.271452 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:38Z","lastTransitionTime":"2025-10-01T09:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.373220 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.373258 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.373269 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.373283 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.373293 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:38Z","lastTransitionTime":"2025-10-01T09:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.476595 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.476632 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.476665 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.476680 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.476690 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:38Z","lastTransitionTime":"2025-10-01T09:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.523526 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.523494 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:37:38 crc kubenswrapper[4787]: E1001 09:37:38.523652 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.523690 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:37:38 crc kubenswrapper[4787]: E1001 09:37:38.523852 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:37:38 crc kubenswrapper[4787]: E1001 09:37:38.523918 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.535573 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebc8a6e2b9c6104b0f773e99d68a62e1b2bf0adc05e6f291cf91e7a8cf1bc121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.547989 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"049825a6-0cec-4adf-8b77-4889bdc4d56f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18044055eb1f6cfd7cb05f330314231b1c563c34a51ba5ed79073988eb563cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc8192c84986df76c9a45deba61aaaa1b0a019fce0000e1c9916c4fe31d1ba1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-84q9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.561610 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.574674 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47825359-aa6e-4315-9a38-685f11f590f7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93d6842d7f8928e5a2b557a70568f5b895e17f9be42d3fadae0b91ba692f5fd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d5cf849bb15a6e2c7abb11fc455dd38be5e20a59158454f27ef69541ff50765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d5cf849bb15a6e2c7abb11fc455dd38be5e20a59158454f27ef69541ff50765\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.579494 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.579546 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.579560 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.579584 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.579598 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:38Z","lastTransitionTime":"2025-10-01T09:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.588489 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.602810 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4990da342a20f6ab1e97774294bd6856c49ca17de21d893a638c13c8e727dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.621354 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b093e96a244fdec4c5cdce62650d4da31c2ec46207a9c08bbd0a28bf286abb52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4664cb1dde25ada88989e81f0319aedc6dbd363f608a50623efaa3eeb8d00763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.633850 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d95f8b1aac879d1746d8a8af7e1d5e672b36c31d5b95c23fee70b237bc3716e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.646906 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7aa5c938897ea621fc7c71c86e6a1dd2acca273e13afe71e9a27fc81f34cf2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.669599 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c68b374f72c2a689eb09480e7ae3e88691daf0cb380a52f655eb1c651a66a281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.681895 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.681944 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.681955 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.681978 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.681992 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:38Z","lastTransitionTime":"2025-10-01T09:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.688149 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:37:21Z\\\",\\\"message\\\":\\\"ift-image-registry/node-ca-5ghjd\\\\nF1001 09:37:21.403568 6768 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:21Z is after 2025-08-24T17:21:41Z]\\\\nI1001 09:37:21.403587 6768 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m\\\\nI1001 09:37:21.403590 6768 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-5ghjd\\\\nI1001 09:37:21.403594 6768 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1001 09:37:21.403595 6768 ovn.go:134] Ensuring zone local for Pod openshift-ovn-k\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:37:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dmsxm_openshift-ovn-kubernetes(eadf2545-886e-4642-b909-704c6cd4134e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.698696 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6zz6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb4a902-9efa-4088-a236-9a634fd085fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k5bxp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k5bxp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6zz6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.713542 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0814bed7-1117-4dff-97ef-a3cd1240e615\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c91542e7254c8289b734b3ffa8e8226f2445c915c264e09c179b99604abad1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:36:27Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1001 09:36:22.478145 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 09:36:22.478910 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3375976971/tls.crt::/tmp/serving-cert-3375976971/tls.key\\\\\\\"\\\\nI1001 09:36:27.943129 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 09:36:27.955126 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 09:36:27.955153 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 09:36:27.955179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 09:36:27.955185 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 09:36:27.963955 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1001 09:36:27.963979 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1001 09:36:27.963986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964008 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 09:36:27.964015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 09:36:27.964018 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 09:36:27.964022 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1001 09:36:27.966112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.726618 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc37698-af76-4818-b216-ffe35af67929\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1bb16edb25b63e957cc0bceb04a9fa9244e68881cb45be7c7d627c5faa74232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://995030421e0cbfeb7baa8ccb574d269dd178f8185bcaa9000fe5e7fc221b1188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d4c78dd607b81a28bcfe0ffa69c02628cf06aff5fd6a574069639c5bac044a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://144123d273f3f1af4fb857d731a16d9891b097c98229a035eb132f360fa8b2f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.738362 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72701c6061105b7bb81545b5f0b035b220b22b2f8009e82809a037870c625685\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.752539 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.777694 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c830596f83deb4fa998fa9af881c7eb1f3673588f444edc9cbf895c19b8248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"message\\\":\\\"2025-10-01T09:36:30+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_54bdebc1-b80c-438d-921a-17ba60e91a86\\\\n2025-10-01T09:36:30+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_54bdebc1-b80c-438d-921a-17ba60e91a86 to /host/opt/cni/bin/\\\\n2025-10-01T09:36:30Z [verbose] multus-daemon started\\\\n2025-10-01T09:36:30Z [verbose] Readiness Indicator file check\\\\n2025-10-01T09:37:15Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.787563 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.787655 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.787678 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.787726 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.787748 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:38Z","lastTransitionTime":"2025-10-01T09:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.794040 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"180a8b2f-41ef-460c-98f2-2914e3b8f9d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f445035e4b6faf9fca8e0dbe8c69668de8825ccebd6ace48f80685c88312b707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f63e01c8f45be048cd6ade60cb51f054fbea42bb51d7b5cd32f71cf1f15ce21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1d95bbc5f966defa47192b7cc53c32f75894f4b51ce8411d7bf828e04c282b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c3cd7c1e19e00dbc9cb8cb8f3f857ae636ff3818c589769f15aa87cb94327ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c3cd7c1e19e00dbc9cb8cb8f3f857ae636ff3818c589769f15aa87cb94327ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:38Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.891395 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.891484 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.891499 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.891544 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.891560 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:38Z","lastTransitionTime":"2025-10-01T09:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.994819 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.994880 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.994899 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.994924 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:38 crc kubenswrapper[4787]: I1001 09:37:38.994940 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:38Z","lastTransitionTime":"2025-10-01T09:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.098530 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.098607 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.098624 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.098659 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.098678 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:39Z","lastTransitionTime":"2025-10-01T09:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.202587 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.202671 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.202690 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.202718 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.202742 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:39Z","lastTransitionTime":"2025-10-01T09:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.306856 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.307038 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.307066 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.307149 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.307171 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:39Z","lastTransitionTime":"2025-10-01T09:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.410803 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.410892 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.410911 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.410946 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.410970 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:39Z","lastTransitionTime":"2025-10-01T09:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.514224 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.514282 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.514296 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.514315 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.514331 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:39Z","lastTransitionTime":"2025-10-01T09:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.522770 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:37:39 crc kubenswrapper[4787]: E1001 09:37:39.522956 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.617116 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.617198 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.617213 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.617231 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.617243 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:39Z","lastTransitionTime":"2025-10-01T09:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.719564 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.719606 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.719619 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.719637 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.719653 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:39Z","lastTransitionTime":"2025-10-01T09:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.822364 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.822409 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.822424 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.822441 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.822453 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:39Z","lastTransitionTime":"2025-10-01T09:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.925992 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.926112 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.926139 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.926172 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:39 crc kubenswrapper[4787]: I1001 09:37:39.926199 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:39Z","lastTransitionTime":"2025-10-01T09:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.027975 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.028048 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.028061 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.028099 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.028114 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:40Z","lastTransitionTime":"2025-10-01T09:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.130718 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.130768 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.130780 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.130798 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.130811 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:40Z","lastTransitionTime":"2025-10-01T09:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.233584 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.233643 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.233661 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.233684 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.233706 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:40Z","lastTransitionTime":"2025-10-01T09:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.335789 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.335848 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.335862 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.335887 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.335902 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:40Z","lastTransitionTime":"2025-10-01T09:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.438141 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.438199 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.438212 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.438228 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.438239 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:40Z","lastTransitionTime":"2025-10-01T09:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.523485 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.523543 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.523620 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:37:40 crc kubenswrapper[4787]: E1001 09:37:40.523733 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:37:40 crc kubenswrapper[4787]: E1001 09:37:40.523913 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:37:40 crc kubenswrapper[4787]: E1001 09:37:40.524351 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.540110 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.540170 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.540184 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.540208 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.540227 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:40Z","lastTransitionTime":"2025-10-01T09:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.644024 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.644093 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.644106 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.644125 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.644137 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:40Z","lastTransitionTime":"2025-10-01T09:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.747882 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.747985 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.747999 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.748028 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.748045 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:40Z","lastTransitionTime":"2025-10-01T09:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.851142 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.851222 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.851235 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.851250 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.851282 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:40Z","lastTransitionTime":"2025-10-01T09:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.954602 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.954649 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.954662 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.954678 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:40 crc kubenswrapper[4787]: I1001 09:37:40.954688 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:40Z","lastTransitionTime":"2025-10-01T09:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.057900 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.057951 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.057962 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.057980 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.057994 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:41Z","lastTransitionTime":"2025-10-01T09:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.161021 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.161164 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.161276 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.161315 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.161374 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:41Z","lastTransitionTime":"2025-10-01T09:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.264299 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.264369 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.264387 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.264411 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.264427 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:41Z","lastTransitionTime":"2025-10-01T09:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.367431 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.367476 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.367486 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.367504 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.367516 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:41Z","lastTransitionTime":"2025-10-01T09:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.470330 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.470390 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.470405 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.470427 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.470441 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:41Z","lastTransitionTime":"2025-10-01T09:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.523213 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:37:41 crc kubenswrapper[4787]: E1001 09:37:41.523401 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.574393 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.574462 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.574484 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.574512 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.574530 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:41Z","lastTransitionTime":"2025-10-01T09:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.678648 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.678741 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.678762 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.678791 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.678810 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:41Z","lastTransitionTime":"2025-10-01T09:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.783863 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.783937 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.783956 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.783985 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.784004 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:41Z","lastTransitionTime":"2025-10-01T09:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.887070 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.887180 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.887201 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.887237 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.887280 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:41Z","lastTransitionTime":"2025-10-01T09:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.990321 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.990392 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.990409 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.990434 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:41 crc kubenswrapper[4787]: I1001 09:37:41.990451 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:41Z","lastTransitionTime":"2025-10-01T09:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.095976 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.096051 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.096064 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.096099 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.096112 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:42Z","lastTransitionTime":"2025-10-01T09:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.198727 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.198772 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.198782 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.198796 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.198806 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:42Z","lastTransitionTime":"2025-10-01T09:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.301975 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.302038 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.302049 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.302095 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.302110 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:42Z","lastTransitionTime":"2025-10-01T09:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.404852 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.404889 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.404900 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.404915 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.404926 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:42Z","lastTransitionTime":"2025-10-01T09:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.507152 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.507193 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.507205 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.507222 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.507234 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:42Z","lastTransitionTime":"2025-10-01T09:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.523606 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.523635 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:37:42 crc kubenswrapper[4787]: E1001 09:37:42.523714 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.523827 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:37:42 crc kubenswrapper[4787]: E1001 09:37:42.523884 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:37:42 crc kubenswrapper[4787]: E1001 09:37:42.524021 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.609643 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.609710 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.609732 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.609759 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.609774 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:42Z","lastTransitionTime":"2025-10-01T09:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.712788 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.712834 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.712842 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.712857 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.712867 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:42Z","lastTransitionTime":"2025-10-01T09:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.815723 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.815817 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.815836 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.815867 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.815890 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:42Z","lastTransitionTime":"2025-10-01T09:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.919745 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.919823 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.919843 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.919874 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:42 crc kubenswrapper[4787]: I1001 09:37:42.919896 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:42Z","lastTransitionTime":"2025-10-01T09:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.022858 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.023022 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.023041 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.023099 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.023120 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:43Z","lastTransitionTime":"2025-10-01T09:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.126702 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.126772 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.126785 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.126811 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.126827 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:43Z","lastTransitionTime":"2025-10-01T09:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.230011 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.230073 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.230128 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.230156 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.230172 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:43Z","lastTransitionTime":"2025-10-01T09:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.333566 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.333663 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.333689 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.333727 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.333756 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:43Z","lastTransitionTime":"2025-10-01T09:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.436294 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.436420 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.436436 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.436455 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.436468 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:43Z","lastTransitionTime":"2025-10-01T09:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.523569 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:37:43 crc kubenswrapper[4787]: E1001 09:37:43.524169 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.539670 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.539723 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.539732 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.539751 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.539764 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:43Z","lastTransitionTime":"2025-10-01T09:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.642388 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.642438 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.642447 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.642461 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.642470 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:43Z","lastTransitionTime":"2025-10-01T09:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.745055 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.745272 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.745290 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.745313 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.745331 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:43Z","lastTransitionTime":"2025-10-01T09:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.849211 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.849270 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.849290 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.849319 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.849337 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:43Z","lastTransitionTime":"2025-10-01T09:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.952661 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.952827 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.952849 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.952878 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:43 crc kubenswrapper[4787]: I1001 09:37:43.952898 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:43Z","lastTransitionTime":"2025-10-01T09:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.055529 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.055601 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.055620 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.055644 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.055665 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:44Z","lastTransitionTime":"2025-10-01T09:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.158411 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.158479 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.158496 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.158520 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.158539 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:44Z","lastTransitionTime":"2025-10-01T09:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.261954 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.262024 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.262041 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.262068 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.262127 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:44Z","lastTransitionTime":"2025-10-01T09:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.365701 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.365754 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.365763 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.365779 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.365794 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:44Z","lastTransitionTime":"2025-10-01T09:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.469168 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.469228 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.469239 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.469256 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.469268 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:44Z","lastTransitionTime":"2025-10-01T09:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.523339 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.523528 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:37:44 crc kubenswrapper[4787]: E1001 09:37:44.523632 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.523732 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:37:44 crc kubenswrapper[4787]: E1001 09:37:44.523847 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:37:44 crc kubenswrapper[4787]: E1001 09:37:44.524019 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.572485 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.572535 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.572545 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.572563 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.572573 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:44Z","lastTransitionTime":"2025-10-01T09:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.675376 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.675435 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.675445 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.675467 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.675481 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:44Z","lastTransitionTime":"2025-10-01T09:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.780393 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.780447 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.780461 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.780481 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.780495 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:44Z","lastTransitionTime":"2025-10-01T09:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.883429 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.883490 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.883499 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.883518 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.883530 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:44Z","lastTransitionTime":"2025-10-01T09:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.986396 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.986481 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.986493 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.986514 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:44 crc kubenswrapper[4787]: I1001 09:37:44.986525 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:44Z","lastTransitionTime":"2025-10-01T09:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.089316 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.089375 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.089390 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.089412 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.089424 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:45Z","lastTransitionTime":"2025-10-01T09:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.192895 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.192948 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.192963 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.192984 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.192996 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:45Z","lastTransitionTime":"2025-10-01T09:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.296252 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.296313 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.296333 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.296360 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.296381 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:45Z","lastTransitionTime":"2025-10-01T09:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.399541 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.399597 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.399608 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.399627 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.399645 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:45Z","lastTransitionTime":"2025-10-01T09:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.502953 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.502995 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.503004 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.503020 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.503030 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:45Z","lastTransitionTime":"2025-10-01T09:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.523757 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:37:45 crc kubenswrapper[4787]: E1001 09:37:45.524191 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.605127 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.605185 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.605196 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.605211 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.605223 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:45Z","lastTransitionTime":"2025-10-01T09:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.708538 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.708612 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.708637 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.708670 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.708693 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:45Z","lastTransitionTime":"2025-10-01T09:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.811623 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.811677 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.811694 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.811721 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.811742 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:45Z","lastTransitionTime":"2025-10-01T09:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.915809 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.915879 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.915900 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.915931 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:45 crc kubenswrapper[4787]: I1001 09:37:45.915956 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:45Z","lastTransitionTime":"2025-10-01T09:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.018911 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.018980 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.019002 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.019031 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.019051 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:46Z","lastTransitionTime":"2025-10-01T09:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.122398 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.122463 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.122474 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.122491 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.122502 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:46Z","lastTransitionTime":"2025-10-01T09:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.176944 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.177033 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.177053 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.177119 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.177149 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:46Z","lastTransitionTime":"2025-10-01T09:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:46 crc kubenswrapper[4787]: E1001 09:37:46.199825 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:46Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.204333 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.204392 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.204405 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.204418 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.204427 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:46Z","lastTransitionTime":"2025-10-01T09:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:46 crc kubenswrapper[4787]: E1001 09:37:46.217561 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:46Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.222965 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.223028 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.223046 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.223071 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.223113 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:46Z","lastTransitionTime":"2025-10-01T09:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:46 crc kubenswrapper[4787]: E1001 09:37:46.237154 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:46Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.241851 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.241930 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.241950 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.241979 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.241999 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:46Z","lastTransitionTime":"2025-10-01T09:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:46 crc kubenswrapper[4787]: E1001 09:37:46.256793 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:46Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.261561 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.261598 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.261610 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.261630 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.261645 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:46Z","lastTransitionTime":"2025-10-01T09:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:46 crc kubenswrapper[4787]: E1001 09:37:46.275589 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:46Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:46 crc kubenswrapper[4787]: E1001 09:37:46.275760 4787 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.277797 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.277844 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.277857 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.277874 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.277885 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:46Z","lastTransitionTime":"2025-10-01T09:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.380493 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.380547 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.380559 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.380578 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.380592 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:46Z","lastTransitionTime":"2025-10-01T09:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.414782 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0eb4a902-9efa-4088-a236-9a634fd085fd-metrics-certs\") pod \"network-metrics-daemon-6zz6g\" (UID: \"0eb4a902-9efa-4088-a236-9a634fd085fd\") " pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:37:46 crc kubenswrapper[4787]: E1001 09:37:46.415011 4787 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 09:37:46 crc kubenswrapper[4787]: E1001 09:37:46.415151 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0eb4a902-9efa-4088-a236-9a634fd085fd-metrics-certs podName:0eb4a902-9efa-4088-a236-9a634fd085fd nodeName:}" failed. No retries permitted until 2025-10-01 09:38:50.415127843 +0000 UTC m=+162.530272020 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0eb4a902-9efa-4088-a236-9a634fd085fd-metrics-certs") pod "network-metrics-daemon-6zz6g" (UID: "0eb4a902-9efa-4088-a236-9a634fd085fd") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.490513 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.490572 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.490586 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.490606 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.490619 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:46Z","lastTransitionTime":"2025-10-01T09:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.523041 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.523145 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.523148 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:37:46 crc kubenswrapper[4787]: E1001 09:37:46.523351 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:37:46 crc kubenswrapper[4787]: E1001 09:37:46.523485 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:37:46 crc kubenswrapper[4787]: E1001 09:37:46.523591 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.592824 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.592872 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.592883 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.592898 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.592908 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:46Z","lastTransitionTime":"2025-10-01T09:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.696299 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.696353 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.696365 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.696384 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.696396 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:46Z","lastTransitionTime":"2025-10-01T09:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.800111 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.800185 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.800206 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.800236 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.800255 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:46Z","lastTransitionTime":"2025-10-01T09:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.903161 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.903202 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.903211 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.903225 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:46 crc kubenswrapper[4787]: I1001 09:37:46.903236 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:46Z","lastTransitionTime":"2025-10-01T09:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.005540 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.005594 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.005602 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.005615 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.005624 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:47Z","lastTransitionTime":"2025-10-01T09:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.107688 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.107748 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.107758 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.107773 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.107784 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:47Z","lastTransitionTime":"2025-10-01T09:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.210169 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.210228 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.210239 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.210258 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.210272 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:47Z","lastTransitionTime":"2025-10-01T09:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.312505 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.312560 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.312571 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.312589 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.312600 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:47Z","lastTransitionTime":"2025-10-01T09:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.415403 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.415463 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.415471 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.415488 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.415497 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:47Z","lastTransitionTime":"2025-10-01T09:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.519067 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.519132 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.519143 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.519162 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.519175 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:47Z","lastTransitionTime":"2025-10-01T09:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.523496 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:37:47 crc kubenswrapper[4787]: E1001 09:37:47.523814 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.524537 4787 scope.go:117] "RemoveContainer" containerID="c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122" Oct 01 09:37:47 crc kubenswrapper[4787]: E1001 09:37:47.524706 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dmsxm_openshift-ovn-kubernetes(eadf2545-886e-4642-b909-704c6cd4134e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" podUID="eadf2545-886e-4642-b909-704c6cd4134e" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.621661 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.621716 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.621729 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.621749 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.621762 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:47Z","lastTransitionTime":"2025-10-01T09:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.724773 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.724816 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.724829 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.724847 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.724862 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:47Z","lastTransitionTime":"2025-10-01T09:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.827006 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.827053 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.827064 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.827104 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.827117 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:47Z","lastTransitionTime":"2025-10-01T09:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.929928 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.929980 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.929992 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.930006 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:47 crc kubenswrapper[4787]: I1001 09:37:47.930015 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:47Z","lastTransitionTime":"2025-10-01T09:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.033218 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.033278 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.033290 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.033305 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.033316 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:48Z","lastTransitionTime":"2025-10-01T09:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.137165 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.137215 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.137227 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.137243 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.137286 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:48Z","lastTransitionTime":"2025-10-01T09:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.239963 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.240038 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.240050 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.240065 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.240096 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:48Z","lastTransitionTime":"2025-10-01T09:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.342833 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.342874 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.342882 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.342898 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.342907 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:48Z","lastTransitionTime":"2025-10-01T09:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.445418 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.445455 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.445466 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.445480 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.445488 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:48Z","lastTransitionTime":"2025-10-01T09:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.523495 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.523573 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.523643 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:37:48 crc kubenswrapper[4787]: E1001 09:37:48.523783 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:37:48 crc kubenswrapper[4787]: E1001 09:37:48.523903 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:37:48 crc kubenswrapper[4787]: E1001 09:37:48.524343 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.547456 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"180a8b2f-41ef-460c-98f2-2914e3b8f9d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f445035e4b6faf9fca8e0dbe8c69668de8825ccebd6ace48f80685c88312b707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f63e01c8f45be048cd6ade60cb51f054fbea42bb51d7b5cd32f71cf1f15ce21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1d95bbc5f966defa47192b7cc53c32f75894f4b51ce8411d7bf828e04c282b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c3cd7c1e19e00dbc9cb8cb8f3f857ae636ff3818c589769f15aa87cb94327ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c3cd7c1e19e00dbc9cb8cb8f3f857ae636ff3818c589769f15aa87cb94327ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.549973 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.550045 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.550069 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.550133 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.550158 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:48Z","lastTransitionTime":"2025-10-01T09:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.570690 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc37698-af76-4818-b216-ffe35af67929\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1bb16edb25b63e957cc0bceb04a9fa9244e68881cb45be7c7d627c5faa74232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://995030421e0cbfeb7baa8ccb574d269dd178f8185bcaa9000fe5e7fc221b1188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8d4c78dd607b81a28bcfe0ffa69c02628cf06aff5fd6a574069639c5bac044a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://144123d273f3f1af4fb857d731a16d9891b097c98229a035eb132f360fa8b2f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.594342 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72701c6061105b7bb81545b5f0b035b220b22b2f8009e82809a037870c625685\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.612278 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.628592 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8tjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53f5ffd2-7649-4c9e-929f-a68187409644\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c830596f83deb4fa998fa9af881c7eb1f3673588f444edc9cbf895c19b8248c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:37:15Z\\\",\\\"message\\\":\\\"2025-10-01T09:36:30+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_54bdebc1-b80c-438d-921a-17ba60e91a86\\\\n2025-10-01T09:36:30+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_54bdebc1-b80c-438d-921a-17ba60e91a86 to /host/opt/cni/bin/\\\\n2025-10-01T09:36:30Z [verbose] multus-daemon started\\\\n2025-10-01T09:36:30Z [verbose] Readiness Indicator file check\\\\n2025-10-01T09:37:15Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:37:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2wcmv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8tjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.651965 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.653145 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.653182 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.653197 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.653220 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.653236 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:48Z","lastTransitionTime":"2025-10-01T09:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.668765 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5ghjd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9b99ece-c0d3-4482-ab68-2672f03b107b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebc8a6e2b9c6104b0f773e99d68a62e1b2bf0adc05e6f291cf91e7a8cf1bc121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bphdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5ghjd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.689583 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"049825a6-0cec-4adf-8b77-4889bdc4d56f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18044055eb1f6cfd7cb05f330314231b1c563c34a51ba5ed79073988eb563cc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc8192c84986df76c9a45deba61aaaa1b0a019fce0000e1c9916c4fe31d1ba1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sqlsn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-84q9m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.706276 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47825359-aa6e-4315-9a38-685f11f590f7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93d6842d7f8928e5a2b557a70568f5b895e17f9be42d3fadae0b91ba692f5fd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d5cf849bb15a6e2c7abb11fc455dd38be5e20a59158454f27ef69541ff50765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d5cf849bb15a6e2c7abb11fc455dd38be5e20a59158454f27ef69541ff50765\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.722213 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.737051 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4990da342a20f6ab1e97774294bd6856c49ca17de21d893a638c13c8e727dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.751888 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b093e96a244fdec4c5cdce62650d4da31c2ec46207a9c08bbd0a28bf286abb52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4664cb1dde25ada88989e81f0319aedc6dbd363f608a50623efaa3eeb8d00763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.756591 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.756696 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.756711 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.756766 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.756783 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:48Z","lastTransitionTime":"2025-10-01T09:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.767410 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa51b883-5f3e-4141-9d4c-37704aac7718\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d95f8b1aac879d1746d8a8af7e1d5e672b36c31d5b95c23fee70b237bc3716e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ppfbw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8vh6d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.785902 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0814bed7-1117-4dff-97ef-a3cd1240e615\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c91542e7254c8289b734b3ffa8e8226f2445c915c264e09c179b99604abad1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:36:27Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1001 09:36:22.478145 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 09:36:22.478910 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3375976971/tls.crt::/tmp/serving-cert-3375976971/tls.key\\\\\\\"\\\\nI1001 09:36:27.943129 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 09:36:27.955126 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 09:36:27.955153 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 09:36:27.955179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 09:36:27.955185 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 09:36:27.963955 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1001 09:36:27.963979 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1001 09:36:27.963986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964008 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 09:36:27.964015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 09:36:27.964018 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 09:36:27.964022 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1001 09:36:27.966112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.805223 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7aa5c938897ea621fc7c71c86e6a1dd2acca273e13afe71e9a27fc81f34cf2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.828604 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kzcns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60771b6a-9398-4987-bb28-a20ceb6cca38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c68b374f72c2a689eb09480e7ae3e88691daf0cb380a52f655eb1c651a66a281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://785665a1ddb89b16f71d8d5e7bacea86ed5e80e22604000fa71b5b982f031878\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58767f74cf601b6e174a361bb41f858c4e12842f28e314c931556eccaafcef26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b5b36b163b4af7f832b70d15a64fead6206acbb9c69d688583980d3fec77b81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee5d303041195225478f984f9efd00c23bce3a0525b239909635a0a2123ff8e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98ea5b20969214b301d2c0de6a94f63740ce94d7c31bcd72b99ee7d6eb95c54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f729cdcc825914b37161e2c9123f69e2e36bd440cb32941fdc4ab611489696b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d7sz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kzcns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.855531 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eadf2545-886e-4642-b909-704c6cd4134e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-01T09:37:21Z\\\",\\\"message\\\":\\\"ift-image-registry/node-ca-5ghjd\\\\nF1001 09:37:21.403568 6768 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:21Z is after 2025-08-24T17:21:41Z]\\\\nI1001 09:37:21.403587 6768 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m\\\\nI1001 09:37:21.403590 6768 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-5ghjd\\\\nI1001 09:37:21.403594 6768 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1001 09:37:21.403595 6768 ovn.go:134] Ensuring zone local for Pod openshift-ovn-k\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:37:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dmsxm_openshift-ovn-kubernetes(eadf2545-886e-4642-b909-704c6cd4134e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z8kgg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dmsxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.859284 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.859404 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.859502 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.859606 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.859693 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:48Z","lastTransitionTime":"2025-10-01T09:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.878925 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6zz6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0eb4a902-9efa-4088-a236-9a634fd085fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k5bxp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k5bxp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6zz6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:48Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.961926 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.962316 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.962581 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.962841 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:48 crc kubenswrapper[4787]: I1001 09:37:48.963151 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:48Z","lastTransitionTime":"2025-10-01T09:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.068047 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.068189 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.068218 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.068255 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.068289 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:49Z","lastTransitionTime":"2025-10-01T09:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.172115 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.172198 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.172209 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.172229 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.172240 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:49Z","lastTransitionTime":"2025-10-01T09:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.276237 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.276578 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.276710 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.276891 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.277012 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:49Z","lastTransitionTime":"2025-10-01T09:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.380786 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.380846 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.380859 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.380879 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.380894 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:49Z","lastTransitionTime":"2025-10-01T09:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.485747 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.485832 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.485851 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.485879 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.485898 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:49Z","lastTransitionTime":"2025-10-01T09:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.523405 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:37:49 crc kubenswrapper[4787]: E1001 09:37:49.524142 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.548815 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.591887 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.591959 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.591973 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.591992 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.592009 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:49Z","lastTransitionTime":"2025-10-01T09:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.695365 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.695435 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.695450 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.695472 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.695488 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:49Z","lastTransitionTime":"2025-10-01T09:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.799002 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.799058 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.799090 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.799111 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.799125 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:49Z","lastTransitionTime":"2025-10-01T09:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.902809 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.902883 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.902904 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.902934 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:49 crc kubenswrapper[4787]: I1001 09:37:49.902956 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:49Z","lastTransitionTime":"2025-10-01T09:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.006210 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.006262 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.006273 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.006287 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.006297 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:50Z","lastTransitionTime":"2025-10-01T09:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.109636 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.109681 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.109692 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.109708 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.109720 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:50Z","lastTransitionTime":"2025-10-01T09:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.212513 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.212565 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.212576 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.212595 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.212608 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:50Z","lastTransitionTime":"2025-10-01T09:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.315190 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.315232 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.315243 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.315260 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.315269 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:50Z","lastTransitionTime":"2025-10-01T09:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.419039 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.419125 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.419213 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.419238 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.419253 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:50Z","lastTransitionTime":"2025-10-01T09:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.521978 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.522038 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.522050 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.522070 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.522105 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:50Z","lastTransitionTime":"2025-10-01T09:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.523467 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.523491 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:37:50 crc kubenswrapper[4787]: E1001 09:37:50.523604 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.523678 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:37:50 crc kubenswrapper[4787]: E1001 09:37:50.523833 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:37:50 crc kubenswrapper[4787]: E1001 09:37:50.524005 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.625277 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.625336 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.625352 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.625370 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.625379 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:50Z","lastTransitionTime":"2025-10-01T09:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.728284 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.728329 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.728341 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.728360 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.728371 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:50Z","lastTransitionTime":"2025-10-01T09:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.831072 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.831153 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.831167 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.831187 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.831200 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:50Z","lastTransitionTime":"2025-10-01T09:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.934699 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.934773 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.934787 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.934811 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:50 crc kubenswrapper[4787]: I1001 09:37:50.934828 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:50Z","lastTransitionTime":"2025-10-01T09:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.038758 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.038853 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.038867 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.038885 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.038897 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:51Z","lastTransitionTime":"2025-10-01T09:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.142410 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.142455 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.142468 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.142485 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.142505 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:51Z","lastTransitionTime":"2025-10-01T09:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.244824 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.244900 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.244919 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.244984 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.245037 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:51Z","lastTransitionTime":"2025-10-01T09:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.349429 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.349483 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.349502 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.349525 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.349555 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:51Z","lastTransitionTime":"2025-10-01T09:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.452514 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.452566 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.452579 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.452602 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.452615 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:51Z","lastTransitionTime":"2025-10-01T09:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.522800 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:37:51 crc kubenswrapper[4787]: E1001 09:37:51.523007 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.555289 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.555340 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.555351 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.555366 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.555377 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:51Z","lastTransitionTime":"2025-10-01T09:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.657664 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.657709 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.657719 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.657732 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.657743 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:51Z","lastTransitionTime":"2025-10-01T09:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.760425 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.760467 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.760478 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.760494 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.760543 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:51Z","lastTransitionTime":"2025-10-01T09:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.863849 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.863937 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.863961 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.863995 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.864018 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:51Z","lastTransitionTime":"2025-10-01T09:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.967928 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.967999 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.968018 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.968047 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:51 crc kubenswrapper[4787]: I1001 09:37:51.968067 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:51Z","lastTransitionTime":"2025-10-01T09:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.071200 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.071296 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.071323 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.071360 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.071392 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:52Z","lastTransitionTime":"2025-10-01T09:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.174723 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.174777 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.174789 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.174807 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.174820 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:52Z","lastTransitionTime":"2025-10-01T09:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.278326 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.278424 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.278459 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.278493 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.278512 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:52Z","lastTransitionTime":"2025-10-01T09:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.381991 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.382055 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.382064 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.382104 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.382120 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:52Z","lastTransitionTime":"2025-10-01T09:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.495847 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.495924 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.495943 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.495972 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.495992 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:52Z","lastTransitionTime":"2025-10-01T09:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.523455 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.523603 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:37:52 crc kubenswrapper[4787]: E1001 09:37:52.523783 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.524397 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:37:52 crc kubenswrapper[4787]: E1001 09:37:52.524583 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:37:52 crc kubenswrapper[4787]: E1001 09:37:52.524786 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.600092 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.600238 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.600269 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.600306 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.600330 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:52Z","lastTransitionTime":"2025-10-01T09:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.703654 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.703721 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.703733 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.703752 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.703764 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:52Z","lastTransitionTime":"2025-10-01T09:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.807166 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.807241 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.807259 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.807287 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.807313 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:52Z","lastTransitionTime":"2025-10-01T09:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.910161 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.910243 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.910261 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.910288 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:52 crc kubenswrapper[4787]: I1001 09:37:52.910313 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:52Z","lastTransitionTime":"2025-10-01T09:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.014448 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.014523 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.014540 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.014570 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.014589 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:53Z","lastTransitionTime":"2025-10-01T09:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.117762 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.117844 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.117857 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.117876 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.117889 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:53Z","lastTransitionTime":"2025-10-01T09:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.221573 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.221663 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.221677 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.221695 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.221796 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:53Z","lastTransitionTime":"2025-10-01T09:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.324496 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.324596 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.324620 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.324651 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.324688 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:53Z","lastTransitionTime":"2025-10-01T09:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.427502 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.427564 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.427586 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.427616 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.427639 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:53Z","lastTransitionTime":"2025-10-01T09:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.523384 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:37:53 crc kubenswrapper[4787]: E1001 09:37:53.523954 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.530699 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.530762 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.530776 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.530797 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.530814 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:53Z","lastTransitionTime":"2025-10-01T09:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.633737 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.633833 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.633854 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.633879 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.633896 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:53Z","lastTransitionTime":"2025-10-01T09:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.736581 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.736645 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.736657 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.736676 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.736688 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:53Z","lastTransitionTime":"2025-10-01T09:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.839313 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.839359 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.839371 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.839387 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.839400 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:53Z","lastTransitionTime":"2025-10-01T09:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.942057 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.942151 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.942165 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.942207 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:53 crc kubenswrapper[4787]: I1001 09:37:53.942224 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:53Z","lastTransitionTime":"2025-10-01T09:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.044686 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.044749 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.044768 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.044794 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.044811 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:54Z","lastTransitionTime":"2025-10-01T09:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.149114 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.149180 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.149193 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.149216 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.149233 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:54Z","lastTransitionTime":"2025-10-01T09:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.252043 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.252132 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.252146 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.252165 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.252178 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:54Z","lastTransitionTime":"2025-10-01T09:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.355009 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.355101 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.355122 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.355145 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.355163 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:54Z","lastTransitionTime":"2025-10-01T09:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.458444 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.458493 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.458509 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.458539 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.458556 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:54Z","lastTransitionTime":"2025-10-01T09:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.523654 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.523743 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.523955 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:37:54 crc kubenswrapper[4787]: E1001 09:37:54.523953 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:37:54 crc kubenswrapper[4787]: E1001 09:37:54.524173 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:37:54 crc kubenswrapper[4787]: E1001 09:37:54.524345 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.563206 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.563290 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.563309 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.563340 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.563361 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:54Z","lastTransitionTime":"2025-10-01T09:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.667143 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.667218 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.667235 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.667269 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.667290 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:54Z","lastTransitionTime":"2025-10-01T09:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.772683 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.772738 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.772750 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.772771 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.772786 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:54Z","lastTransitionTime":"2025-10-01T09:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.876052 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.876155 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.876169 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.876191 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.876205 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:54Z","lastTransitionTime":"2025-10-01T09:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.979313 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.979389 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.979409 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.979440 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:54 crc kubenswrapper[4787]: I1001 09:37:54.979461 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:54Z","lastTransitionTime":"2025-10-01T09:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.082128 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.082172 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.082183 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.082202 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.082219 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:55Z","lastTransitionTime":"2025-10-01T09:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.190256 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.190326 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.190345 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.190374 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.190392 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:55Z","lastTransitionTime":"2025-10-01T09:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.294688 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.294753 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.294773 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.294797 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.294815 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:55Z","lastTransitionTime":"2025-10-01T09:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.398197 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.398262 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.398286 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.398319 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.398342 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:55Z","lastTransitionTime":"2025-10-01T09:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.501025 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.501148 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.501168 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.501189 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.501204 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:55Z","lastTransitionTime":"2025-10-01T09:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.522713 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:37:55 crc kubenswrapper[4787]: E1001 09:37:55.522883 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.603860 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.603915 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.603928 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.603947 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.603961 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:55Z","lastTransitionTime":"2025-10-01T09:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.707287 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.707323 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.707335 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.707351 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.707363 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:55Z","lastTransitionTime":"2025-10-01T09:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.810704 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.810783 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.810801 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.810828 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.810845 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:55Z","lastTransitionTime":"2025-10-01T09:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.913744 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.913804 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.913821 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.913850 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:55 crc kubenswrapper[4787]: I1001 09:37:55.913871 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:55Z","lastTransitionTime":"2025-10-01T09:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.018274 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.018386 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.018407 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.018433 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.018450 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:56Z","lastTransitionTime":"2025-10-01T09:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.120876 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.120958 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.120985 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.121013 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.121033 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:56Z","lastTransitionTime":"2025-10-01T09:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.223895 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.223959 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.223976 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.224000 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.224016 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:56Z","lastTransitionTime":"2025-10-01T09:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.326560 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.326612 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.326628 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.326674 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.326685 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:56Z","lastTransitionTime":"2025-10-01T09:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.337729 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.337806 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.337833 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.337862 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.337884 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:56Z","lastTransitionTime":"2025-10-01T09:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:56 crc kubenswrapper[4787]: E1001 09:37:56.358751 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.364602 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.364668 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.364694 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.364724 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.364762 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:56Z","lastTransitionTime":"2025-10-01T09:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:56 crc kubenswrapper[4787]: E1001 09:37:56.379048 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.385872 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.385934 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.385953 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.385982 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.386004 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:56Z","lastTransitionTime":"2025-10-01T09:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:56 crc kubenswrapper[4787]: E1001 09:37:56.405548 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.411009 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.411069 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.411116 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.411146 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.411172 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:56Z","lastTransitionTime":"2025-10-01T09:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:56 crc kubenswrapper[4787]: E1001 09:37:56.430186 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.435150 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.435212 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.435226 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.435252 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.435267 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:56Z","lastTransitionTime":"2025-10-01T09:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:56 crc kubenswrapper[4787]: E1001 09:37:56.453994 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-01T09:37:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-01T09:37:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7f63d222-e40c-4123-900b-74cfec920eac\\\",\\\"systemUUID\\\":\\\"869e21e0-7910-4427-aeba-5df675f5c14f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:56Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:56 crc kubenswrapper[4787]: E1001 09:37:56.454187 4787 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.456265 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.456310 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.456324 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.456344 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.456359 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:56Z","lastTransitionTime":"2025-10-01T09:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.523399 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.523502 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:37:56 crc kubenswrapper[4787]: E1001 09:37:56.523549 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.523690 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:37:56 crc kubenswrapper[4787]: E1001 09:37:56.523759 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:37:56 crc kubenswrapper[4787]: E1001 09:37:56.523907 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.559216 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.559259 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.559269 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.559283 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.559295 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:56Z","lastTransitionTime":"2025-10-01T09:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.662710 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.662792 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.662811 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.662841 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.662862 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:56Z","lastTransitionTime":"2025-10-01T09:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.767063 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.767165 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.767181 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.767208 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.767227 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:56Z","lastTransitionTime":"2025-10-01T09:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.870297 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.870381 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.870401 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.870432 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.870453 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:56Z","lastTransitionTime":"2025-10-01T09:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.973597 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.973645 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.973658 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.973677 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:56 crc kubenswrapper[4787]: I1001 09:37:56.973689 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:56Z","lastTransitionTime":"2025-10-01T09:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.077416 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.077503 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.077527 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.077559 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.077583 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:57Z","lastTransitionTime":"2025-10-01T09:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.180598 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.180675 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.180698 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.180732 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.180760 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:57Z","lastTransitionTime":"2025-10-01T09:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.284728 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.284806 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.284831 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.284868 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.284893 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:57Z","lastTransitionTime":"2025-10-01T09:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.388288 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.388390 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.388413 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.388446 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.388468 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:57Z","lastTransitionTime":"2025-10-01T09:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.492009 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.492095 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.492114 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.492142 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.492161 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:57Z","lastTransitionTime":"2025-10-01T09:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.523796 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:37:57 crc kubenswrapper[4787]: E1001 09:37:57.524251 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.594898 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.595006 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.595024 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.595052 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.595071 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:57Z","lastTransitionTime":"2025-10-01T09:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.698698 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.698838 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.698864 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.698896 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.698918 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:57Z","lastTransitionTime":"2025-10-01T09:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.800970 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.801039 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.801060 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.801116 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.801137 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:57Z","lastTransitionTime":"2025-10-01T09:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.903256 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.903313 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.903327 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.903345 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:57 crc kubenswrapper[4787]: I1001 09:37:57.903359 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:57Z","lastTransitionTime":"2025-10-01T09:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.005796 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.005842 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.005853 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.005869 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.005882 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:58Z","lastTransitionTime":"2025-10-01T09:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.108407 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.108480 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.108512 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.108548 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.108569 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:58Z","lastTransitionTime":"2025-10-01T09:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.211794 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.211862 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.211881 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.211906 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.211924 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:58Z","lastTransitionTime":"2025-10-01T09:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.315030 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.315097 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.315113 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.315132 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.315144 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:58Z","lastTransitionTime":"2025-10-01T09:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.418408 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.418482 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.418506 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.418540 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.418563 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:58Z","lastTransitionTime":"2025-10-01T09:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.520407 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.520440 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.520450 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.520462 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.520471 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:58Z","lastTransitionTime":"2025-10-01T09:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.522976 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.523225 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.523265 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:37:58 crc kubenswrapper[4787]: E1001 09:37:58.523372 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:37:58 crc kubenswrapper[4787]: E1001 09:37:58.523526 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:37:58 crc kubenswrapper[4787]: E1001 09:37:58.523635 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.538597 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0814bed7-1117-4dff-97ef-a3cd1240e615\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b87441bd0a43f4473bf5e5494edc0046df0504973093d2e074fd205d438e283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e56def21c50f83e86eb607170e73cad25b255660c179558b6f73793bb3a42644\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca049374076b12b69fcf07f6863b1d0e63f0507539799fd1d46f0683bd1d5c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c91542e7254c8289b734b3ffa8e8226f2445c915c264e09c179b99604abad1a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://149cddc34a611604ea5da19cdaaf329db79f128e454f6bde901b8a5b4ebdbb4c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-01T09:36:27Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1001 09:36:22.478145 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1001 09:36:22.478910 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3375976971/tls.crt::/tmp/serving-cert-3375976971/tls.key\\\\\\\"\\\\nI1001 09:36:27.943129 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1001 09:36:27.955126 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1001 09:36:27.955153 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1001 09:36:27.955179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1001 09:36:27.955185 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1001 09:36:27.963955 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1001 09:36:27.963979 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1001 09:36:27.963986 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964008 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1001 09:36:27.964012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1001 09:36:27.964015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1001 09:36:27.964018 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1001 09:36:27.964022 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1001 09:36:27.966112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d02b4c02e1045dadad35bd1ea8f6f9f7ce29705386ce9fcbdede2521e375b730\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:11Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a889d4969981e808e937a815bf9b4e75129050e0bfa9e4fb8b71bbe8ed195d6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-01T09:36:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-01T09:36:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:08Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:58Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.548037 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cfzj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bba21c1-f38f-418b-add8-bff4953fe9a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-01T09:36:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7aa5c938897ea621fc7c71c86e6a1dd2acca273e13afe71e9a27fc81f34cf2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-01T09:36:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9mvlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-01T09:36:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cfzj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-01T09:37:58Z is after 2025-08-24T17:21:41Z" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.580616 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-kzcns" podStartSLOduration=90.580563146 podStartE2EDuration="1m30.580563146s" podCreationTimestamp="2025-10-01 09:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:37:58.579199472 +0000 UTC m=+110.694343639" watchObservedRunningTime="2025-10-01 09:37:58.580563146 +0000 UTC m=+110.695707303" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.622057 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.622135 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.622150 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.622188 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.622201 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:58Z","lastTransitionTime":"2025-10-01T09:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.624765 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=59.62475347 podStartE2EDuration="59.62475347s" podCreationTimestamp="2025-10-01 09:36:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:37:58.624180406 +0000 UTC m=+110.739324583" watchObservedRunningTime="2025-10-01 09:37:58.62475347 +0000 UTC m=+110.739897627" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.640899 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=85.640884519 podStartE2EDuration="1m25.640884519s" podCreationTimestamp="2025-10-01 09:36:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:37:58.638737297 +0000 UTC m=+110.753881454" watchObservedRunningTime="2025-10-01 09:37:58.640884519 +0000 UTC m=+110.756028676" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.686813 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-z8tjz" podStartSLOduration=90.686795726 podStartE2EDuration="1m30.686795726s" podCreationTimestamp="2025-10-01 09:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:37:58.685393761 +0000 UTC m=+110.800537938" watchObservedRunningTime="2025-10-01 09:37:58.686795726 +0000 UTC m=+110.801939883" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.719567 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-5ghjd" podStartSLOduration=91.719548546 podStartE2EDuration="1m31.719548546s" podCreationTimestamp="2025-10-01 09:36:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:37:58.706646267 +0000 UTC m=+110.821790424" watchObservedRunningTime="2025-10-01 09:37:58.719548546 +0000 UTC m=+110.834692703" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.724223 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.724262 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.724275 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.724290 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.724318 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:58Z","lastTransitionTime":"2025-10-01T09:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.783658 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-84q9m" podStartSLOduration=90.783633853 podStartE2EDuration="1m30.783633853s" podCreationTimestamp="2025-10-01 09:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:37:58.720590532 +0000 UTC m=+110.835734699" watchObservedRunningTime="2025-10-01 09:37:58.783633853 +0000 UTC m=+110.898778010" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.784662 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=9.784655828 podStartE2EDuration="9.784655828s" podCreationTimestamp="2025-10-01 09:37:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:37:58.783281844 +0000 UTC m=+110.898426011" watchObservedRunningTime="2025-10-01 09:37:58.784655828 +0000 UTC m=+110.899799985" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.801491 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=41.801462334 podStartE2EDuration="41.801462334s" podCreationTimestamp="2025-10-01 09:37:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:37:58.800904601 +0000 UTC m=+110.916048758" watchObservedRunningTime="2025-10-01 09:37:58.801462334 +0000 UTC m=+110.916606491" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.827028 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.827125 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.827138 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.827164 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.827177 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:58Z","lastTransitionTime":"2025-10-01T09:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.858300 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podStartSLOduration=91.858271471 podStartE2EDuration="1m31.858271471s" podCreationTimestamp="2025-10-01 09:36:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:37:58.858116156 +0000 UTC m=+110.973260343" watchObservedRunningTime="2025-10-01 09:37:58.858271471 +0000 UTC m=+110.973415628" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.929567 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.929594 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.929601 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.929615 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:58 crc kubenswrapper[4787]: I1001 09:37:58.929625 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:58Z","lastTransitionTime":"2025-10-01T09:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.031815 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.031849 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.031857 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.031869 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.031878 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:59Z","lastTransitionTime":"2025-10-01T09:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.135112 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.135187 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.135205 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.135233 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.135252 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:59Z","lastTransitionTime":"2025-10-01T09:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.238345 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.238379 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.238389 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.238404 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.238413 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:59Z","lastTransitionTime":"2025-10-01T09:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.341994 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.342133 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.342160 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.342199 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.342227 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:59Z","lastTransitionTime":"2025-10-01T09:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.445669 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.445721 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.445736 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.445760 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.445772 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:59Z","lastTransitionTime":"2025-10-01T09:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.523662 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:37:59 crc kubenswrapper[4787]: E1001 09:37:59.523989 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.548572 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.548634 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.548653 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.548677 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.548695 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:59Z","lastTransitionTime":"2025-10-01T09:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.651973 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.652027 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.652039 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.652059 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.652071 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:59Z","lastTransitionTime":"2025-10-01T09:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.754507 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.754546 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.754554 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.754567 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.754576 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:59Z","lastTransitionTime":"2025-10-01T09:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.857048 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.857108 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.857119 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.857137 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.857148 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:59Z","lastTransitionTime":"2025-10-01T09:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.959874 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.959958 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.959978 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.960009 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:37:59 crc kubenswrapper[4787]: I1001 09:37:59.960028 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:37:59Z","lastTransitionTime":"2025-10-01T09:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.062735 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.062785 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.062796 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.062812 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.062823 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:00Z","lastTransitionTime":"2025-10-01T09:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.166446 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.166516 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.166532 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.166556 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.166571 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:00Z","lastTransitionTime":"2025-10-01T09:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.269617 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.269704 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.269728 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.269758 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.269782 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:00Z","lastTransitionTime":"2025-10-01T09:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.372691 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.372765 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.372787 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.372819 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.372846 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:00Z","lastTransitionTime":"2025-10-01T09:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.475734 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.475814 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.475839 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.475867 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.475889 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:00Z","lastTransitionTime":"2025-10-01T09:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.523774 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.523802 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.523921 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:38:00 crc kubenswrapper[4787]: E1001 09:38:00.524071 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:38:00 crc kubenswrapper[4787]: E1001 09:38:00.524740 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:38:00 crc kubenswrapper[4787]: E1001 09:38:00.525123 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.525182 4787 scope.go:117] "RemoveContainer" containerID="c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122" Oct 01 09:38:00 crc kubenswrapper[4787]: E1001 09:38:00.525421 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dmsxm_openshift-ovn-kubernetes(eadf2545-886e-4642-b909-704c6cd4134e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" podUID="eadf2545-886e-4642-b909-704c6cd4134e" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.578706 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.578767 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.578790 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.578819 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.578840 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:00Z","lastTransitionTime":"2025-10-01T09:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.681105 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.681196 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.681217 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.681240 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.681257 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:00Z","lastTransitionTime":"2025-10-01T09:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.785129 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.785173 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.785203 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.785219 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.785229 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:00Z","lastTransitionTime":"2025-10-01T09:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.888540 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.888573 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.888582 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.888595 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.888603 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:00Z","lastTransitionTime":"2025-10-01T09:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.992853 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.992892 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.992901 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.992933 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:00 crc kubenswrapper[4787]: I1001 09:38:00.992942 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:00Z","lastTransitionTime":"2025-10-01T09:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.095880 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.095929 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.095938 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.095954 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.095965 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:01Z","lastTransitionTime":"2025-10-01T09:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.198798 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.198833 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.198841 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.198855 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.198863 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:01Z","lastTransitionTime":"2025-10-01T09:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.302017 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.302059 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.302069 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.302124 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.302135 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:01Z","lastTransitionTime":"2025-10-01T09:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.404481 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.404530 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.404542 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.404558 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.404572 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:01Z","lastTransitionTime":"2025-10-01T09:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.507250 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.507294 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.507303 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.507318 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.507327 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:01Z","lastTransitionTime":"2025-10-01T09:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.522828 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:38:01 crc kubenswrapper[4787]: E1001 09:38:01.523005 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.610591 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.610654 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.610665 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.610682 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.610700 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:01Z","lastTransitionTime":"2025-10-01T09:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.714679 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.714726 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.714735 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.714755 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.714766 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:01Z","lastTransitionTime":"2025-10-01T09:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.817675 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.817713 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.817909 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.817925 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.817937 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:01Z","lastTransitionTime":"2025-10-01T09:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.921629 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.921717 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.921752 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.921784 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:01 crc kubenswrapper[4787]: I1001 09:38:01.921806 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:01Z","lastTransitionTime":"2025-10-01T09:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.024383 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.024437 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.024454 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.024477 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.024496 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:02Z","lastTransitionTime":"2025-10-01T09:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.096791 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z8tjz_53f5ffd2-7649-4c9e-929f-a68187409644/kube-multus/1.log" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.097903 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z8tjz_53f5ffd2-7649-4c9e-929f-a68187409644/kube-multus/0.log" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.098013 4787 generic.go:334] "Generic (PLEG): container finished" podID="53f5ffd2-7649-4c9e-929f-a68187409644" containerID="8c830596f83deb4fa998fa9af881c7eb1f3673588f444edc9cbf895c19b8248c" exitCode=1 Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.098175 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-z8tjz" event={"ID":"53f5ffd2-7649-4c9e-929f-a68187409644","Type":"ContainerDied","Data":"8c830596f83deb4fa998fa9af881c7eb1f3673588f444edc9cbf895c19b8248c"} Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.098391 4787 scope.go:117] "RemoveContainer" containerID="cecc8e29630fbd39d5da8cdcbafd00a589081305f55f5ab81863b3c0dec56885" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.099029 4787 scope.go:117] "RemoveContainer" containerID="8c830596f83deb4fa998fa9af881c7eb1f3673588f444edc9cbf895c19b8248c" Oct 01 09:38:02 crc kubenswrapper[4787]: E1001 09:38:02.099334 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-z8tjz_openshift-multus(53f5ffd2-7649-4c9e-929f-a68187409644)\"" pod="openshift-multus/multus-z8tjz" podUID="53f5ffd2-7649-4c9e-929f-a68187409644" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.127427 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.128166 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.128200 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.128229 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.128251 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:02Z","lastTransitionTime":"2025-10-01T09:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.130415 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=94.130390614 podStartE2EDuration="1m34.130390614s" podCreationTimestamp="2025-10-01 09:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:02.128875017 +0000 UTC m=+114.244019204" watchObservedRunningTime="2025-10-01 09:38:02.130390614 +0000 UTC m=+114.245534811" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.152261 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-cfzj5" podStartSLOduration=95.152230105 podStartE2EDuration="1m35.152230105s" podCreationTimestamp="2025-10-01 09:36:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:02.149028966 +0000 UTC m=+114.264173173" watchObservedRunningTime="2025-10-01 09:38:02.152230105 +0000 UTC m=+114.267374352" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.231108 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.231160 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.231175 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.231196 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.231211 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:02Z","lastTransitionTime":"2025-10-01T09:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.333705 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.333793 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.333802 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.333821 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.333869 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:02Z","lastTransitionTime":"2025-10-01T09:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.436117 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.436367 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.436499 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.436611 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.436711 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:02Z","lastTransitionTime":"2025-10-01T09:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.523381 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.523428 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:38:02 crc kubenswrapper[4787]: E1001 09:38:02.523513 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.523626 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:38:02 crc kubenswrapper[4787]: E1001 09:38:02.523894 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:38:02 crc kubenswrapper[4787]: E1001 09:38:02.524003 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.538704 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.538744 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.538760 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.538790 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.538802 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:02Z","lastTransitionTime":"2025-10-01T09:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.641334 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.641381 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.641391 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.641404 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.641414 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:02Z","lastTransitionTime":"2025-10-01T09:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.743933 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.743975 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.743986 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.744003 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.744017 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:02Z","lastTransitionTime":"2025-10-01T09:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.846802 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.846844 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.846852 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.846866 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.846875 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:02Z","lastTransitionTime":"2025-10-01T09:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.949460 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.949513 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.949525 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.949543 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:02 crc kubenswrapper[4787]: I1001 09:38:02.949554 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:02Z","lastTransitionTime":"2025-10-01T09:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.052216 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.052291 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.052313 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.052332 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.052346 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:03Z","lastTransitionTime":"2025-10-01T09:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.102554 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z8tjz_53f5ffd2-7649-4c9e-929f-a68187409644/kube-multus/1.log" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.154185 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.154213 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.154223 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.154237 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.154246 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:03Z","lastTransitionTime":"2025-10-01T09:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.256598 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.256857 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.256959 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.257055 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.257221 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:03Z","lastTransitionTime":"2025-10-01T09:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.359491 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.359959 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.360053 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.360181 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.360294 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:03Z","lastTransitionTime":"2025-10-01T09:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.462736 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.463017 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.463118 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.463197 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.463260 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:03Z","lastTransitionTime":"2025-10-01T09:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.523572 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:38:03 crc kubenswrapper[4787]: E1001 09:38:03.523695 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.565295 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.565562 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.565638 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.565748 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.565833 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:03Z","lastTransitionTime":"2025-10-01T09:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.667361 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.667578 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.667672 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.667737 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.667797 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:03Z","lastTransitionTime":"2025-10-01T09:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.770254 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.770282 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.770290 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.770329 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.770340 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:03Z","lastTransitionTime":"2025-10-01T09:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.872036 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.872090 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.872100 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.872114 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.872126 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:03Z","lastTransitionTime":"2025-10-01T09:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.974168 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.974196 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.974205 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.974218 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:03 crc kubenswrapper[4787]: I1001 09:38:03.974228 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:03Z","lastTransitionTime":"2025-10-01T09:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.076743 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.076778 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.076787 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.076802 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.076812 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:04Z","lastTransitionTime":"2025-10-01T09:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.180043 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.180103 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.180113 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.180127 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.180155 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:04Z","lastTransitionTime":"2025-10-01T09:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.282413 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.282453 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.282466 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.282484 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.282495 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:04Z","lastTransitionTime":"2025-10-01T09:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.385127 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.385196 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.385207 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.385247 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.385261 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:04Z","lastTransitionTime":"2025-10-01T09:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.487954 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.487988 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.488013 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.488026 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.488037 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:04Z","lastTransitionTime":"2025-10-01T09:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.522733 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.522798 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:38:04 crc kubenswrapper[4787]: E1001 09:38:04.522869 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.522952 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:38:04 crc kubenswrapper[4787]: E1001 09:38:04.523037 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:38:04 crc kubenswrapper[4787]: E1001 09:38:04.523127 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.590536 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.590572 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.590584 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.590600 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.590611 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:04Z","lastTransitionTime":"2025-10-01T09:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.693263 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.693322 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.693330 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.693343 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.693383 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:04Z","lastTransitionTime":"2025-10-01T09:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.795606 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.795662 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.795670 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.795683 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.795692 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:04Z","lastTransitionTime":"2025-10-01T09:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.898069 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.898126 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.898136 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.898153 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:04 crc kubenswrapper[4787]: I1001 09:38:04.898165 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:04Z","lastTransitionTime":"2025-10-01T09:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.000477 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.000562 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.000575 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.000609 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.000619 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:05Z","lastTransitionTime":"2025-10-01T09:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.102598 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.102637 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.102649 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.102665 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.102675 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:05Z","lastTransitionTime":"2025-10-01T09:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.204569 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.204606 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.204617 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.204631 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.204643 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:05Z","lastTransitionTime":"2025-10-01T09:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.306885 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.306930 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.306941 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.306956 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.306969 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:05Z","lastTransitionTime":"2025-10-01T09:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.409868 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.409917 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.409954 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.409974 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.409984 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:05Z","lastTransitionTime":"2025-10-01T09:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.512253 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.512337 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.512382 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.512399 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.512412 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:05Z","lastTransitionTime":"2025-10-01T09:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.522957 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:38:05 crc kubenswrapper[4787]: E1001 09:38:05.523132 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.614800 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.614844 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.614854 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.614870 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.614882 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:05Z","lastTransitionTime":"2025-10-01T09:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.716704 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.716757 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.716766 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.716782 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.716791 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:05Z","lastTransitionTime":"2025-10-01T09:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.818916 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.818960 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.818986 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.819001 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.819014 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:05Z","lastTransitionTime":"2025-10-01T09:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.925812 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.925872 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.925882 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.925894 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:05 crc kubenswrapper[4787]: I1001 09:38:05.925903 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:05Z","lastTransitionTime":"2025-10-01T09:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.027557 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.027602 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.027611 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.027626 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.027635 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:06Z","lastTransitionTime":"2025-10-01T09:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.129181 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.129222 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.129231 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.129263 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.129273 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:06Z","lastTransitionTime":"2025-10-01T09:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.231176 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.231222 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.231242 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.231259 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.231270 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:06Z","lastTransitionTime":"2025-10-01T09:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.333413 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.333446 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.333454 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.333466 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.333476 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:06Z","lastTransitionTime":"2025-10-01T09:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.434936 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.434992 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.435009 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.435030 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.435048 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:06Z","lastTransitionTime":"2025-10-01T09:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.505792 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.505839 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.505852 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.505870 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.505883 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-01T09:38:06Z","lastTransitionTime":"2025-10-01T09:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.522889 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.523017 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.522909 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:38:06 crc kubenswrapper[4787]: E1001 09:38:06.523133 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:38:06 crc kubenswrapper[4787]: E1001 09:38:06.523447 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:38:06 crc kubenswrapper[4787]: E1001 09:38:06.523502 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.552036 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-w59kp"] Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.552476 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w59kp" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.554523 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.554703 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.554781 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.556272 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.631699 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9d0de636-cbe3-486b-b81f-93ce1481cab7-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-w59kp\" (UID: \"9d0de636-cbe3-486b-b81f-93ce1481cab7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w59kp" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.631736 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9d0de636-cbe3-486b-b81f-93ce1481cab7-service-ca\") pod \"cluster-version-operator-5c965bbfc6-w59kp\" (UID: \"9d0de636-cbe3-486b-b81f-93ce1481cab7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w59kp" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.631758 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/9d0de636-cbe3-486b-b81f-93ce1481cab7-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-w59kp\" (UID: \"9d0de636-cbe3-486b-b81f-93ce1481cab7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w59kp" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.631782 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d0de636-cbe3-486b-b81f-93ce1481cab7-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-w59kp\" (UID: \"9d0de636-cbe3-486b-b81f-93ce1481cab7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w59kp" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.631866 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/9d0de636-cbe3-486b-b81f-93ce1481cab7-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-w59kp\" (UID: \"9d0de636-cbe3-486b-b81f-93ce1481cab7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w59kp" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.733013 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9d0de636-cbe3-486b-b81f-93ce1481cab7-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-w59kp\" (UID: \"9d0de636-cbe3-486b-b81f-93ce1481cab7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w59kp" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.733050 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9d0de636-cbe3-486b-b81f-93ce1481cab7-service-ca\") pod \"cluster-version-operator-5c965bbfc6-w59kp\" (UID: \"9d0de636-cbe3-486b-b81f-93ce1481cab7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w59kp" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.733102 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/9d0de636-cbe3-486b-b81f-93ce1481cab7-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-w59kp\" (UID: \"9d0de636-cbe3-486b-b81f-93ce1481cab7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w59kp" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.733128 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d0de636-cbe3-486b-b81f-93ce1481cab7-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-w59kp\" (UID: \"9d0de636-cbe3-486b-b81f-93ce1481cab7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w59kp" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.733161 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/9d0de636-cbe3-486b-b81f-93ce1481cab7-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-w59kp\" (UID: \"9d0de636-cbe3-486b-b81f-93ce1481cab7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w59kp" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.733218 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/9d0de636-cbe3-486b-b81f-93ce1481cab7-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-w59kp\" (UID: \"9d0de636-cbe3-486b-b81f-93ce1481cab7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w59kp" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.733283 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/9d0de636-cbe3-486b-b81f-93ce1481cab7-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-w59kp\" (UID: \"9d0de636-cbe3-486b-b81f-93ce1481cab7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w59kp" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.734902 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9d0de636-cbe3-486b-b81f-93ce1481cab7-service-ca\") pod \"cluster-version-operator-5c965bbfc6-w59kp\" (UID: \"9d0de636-cbe3-486b-b81f-93ce1481cab7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w59kp" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.739915 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d0de636-cbe3-486b-b81f-93ce1481cab7-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-w59kp\" (UID: \"9d0de636-cbe3-486b-b81f-93ce1481cab7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w59kp" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.748389 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9d0de636-cbe3-486b-b81f-93ce1481cab7-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-w59kp\" (UID: \"9d0de636-cbe3-486b-b81f-93ce1481cab7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w59kp" Oct 01 09:38:06 crc kubenswrapper[4787]: I1001 09:38:06.869602 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w59kp" Oct 01 09:38:07 crc kubenswrapper[4787]: I1001 09:38:07.115135 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w59kp" event={"ID":"9d0de636-cbe3-486b-b81f-93ce1481cab7","Type":"ContainerStarted","Data":"44f103f94c384513bcfab8ff630c8122f3c4ad387ff6f5f67b67463d24810559"} Oct 01 09:38:07 crc kubenswrapper[4787]: I1001 09:38:07.115181 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w59kp" event={"ID":"9d0de636-cbe3-486b-b81f-93ce1481cab7","Type":"ContainerStarted","Data":"908fa93f7544fd4d3c8aa22e8856bc3e4ac36bd04ac4599b40f3f118f06d38c3"} Oct 01 09:38:07 crc kubenswrapper[4787]: I1001 09:38:07.131467 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w59kp" podStartSLOduration=99.131448244 podStartE2EDuration="1m39.131448244s" podCreationTimestamp="2025-10-01 09:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:07.127865956 +0000 UTC m=+119.243010153" watchObservedRunningTime="2025-10-01 09:38:07.131448244 +0000 UTC m=+119.246592391" Oct 01 09:38:07 crc kubenswrapper[4787]: I1001 09:38:07.523362 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:38:07 crc kubenswrapper[4787]: E1001 09:38:07.523492 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:38:08 crc kubenswrapper[4787]: E1001 09:38:08.489886 4787 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 01 09:38:08 crc kubenswrapper[4787]: I1001 09:38:08.522872 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:38:08 crc kubenswrapper[4787]: I1001 09:38:08.522979 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:38:08 crc kubenswrapper[4787]: E1001 09:38:08.523730 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:38:08 crc kubenswrapper[4787]: I1001 09:38:08.523753 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:38:08 crc kubenswrapper[4787]: E1001 09:38:08.523917 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:38:08 crc kubenswrapper[4787]: E1001 09:38:08.524000 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:38:08 crc kubenswrapper[4787]: E1001 09:38:08.636187 4787 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 01 09:38:09 crc kubenswrapper[4787]: I1001 09:38:09.523351 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:38:09 crc kubenswrapper[4787]: E1001 09:38:09.523496 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:38:10 crc kubenswrapper[4787]: I1001 09:38:10.523566 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:38:10 crc kubenswrapper[4787]: I1001 09:38:10.523566 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:38:10 crc kubenswrapper[4787]: I1001 09:38:10.523748 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:38:10 crc kubenswrapper[4787]: E1001 09:38:10.523804 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:38:10 crc kubenswrapper[4787]: E1001 09:38:10.523713 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:38:10 crc kubenswrapper[4787]: E1001 09:38:10.523959 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:38:11 crc kubenswrapper[4787]: I1001 09:38:11.523660 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:38:11 crc kubenswrapper[4787]: E1001 09:38:11.524455 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:38:12 crc kubenswrapper[4787]: I1001 09:38:12.523978 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:38:12 crc kubenswrapper[4787]: E1001 09:38:12.524867 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:38:12 crc kubenswrapper[4787]: I1001 09:38:12.524621 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:38:12 crc kubenswrapper[4787]: I1001 09:38:12.525012 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:38:12 crc kubenswrapper[4787]: E1001 09:38:12.525283 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:38:12 crc kubenswrapper[4787]: E1001 09:38:12.525399 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:38:13 crc kubenswrapper[4787]: I1001 09:38:13.522849 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:38:13 crc kubenswrapper[4787]: E1001 09:38:13.523221 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:38:13 crc kubenswrapper[4787]: I1001 09:38:13.525531 4787 scope.go:117] "RemoveContainer" containerID="c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122" Oct 01 09:38:13 crc kubenswrapper[4787]: E1001 09:38:13.638050 4787 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 01 09:38:14 crc kubenswrapper[4787]: I1001 09:38:14.146964 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dmsxm_eadf2545-886e-4642-b909-704c6cd4134e/ovnkube-controller/3.log" Oct 01 09:38:14 crc kubenswrapper[4787]: I1001 09:38:14.150580 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" event={"ID":"eadf2545-886e-4642-b909-704c6cd4134e","Type":"ContainerStarted","Data":"81f37f28240c3a86515abe10685e1a875a8cafa2df45d92b131677f3b1eacdac"} Oct 01 09:38:14 crc kubenswrapper[4787]: I1001 09:38:14.150977 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:38:14 crc kubenswrapper[4787]: I1001 09:38:14.189318 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" podStartSLOduration=106.189299065 podStartE2EDuration="1m46.189299065s" podCreationTimestamp="2025-10-01 09:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:14.189242343 +0000 UTC m=+126.304386550" watchObservedRunningTime="2025-10-01 09:38:14.189299065 +0000 UTC m=+126.304443222" Oct 01 09:38:14 crc kubenswrapper[4787]: I1001 09:38:14.424528 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-6zz6g"] Oct 01 09:38:14 crc kubenswrapper[4787]: I1001 09:38:14.424625 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:38:14 crc kubenswrapper[4787]: E1001 09:38:14.424700 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:38:14 crc kubenswrapper[4787]: I1001 09:38:14.523174 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:38:14 crc kubenswrapper[4787]: I1001 09:38:14.523223 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:38:14 crc kubenswrapper[4787]: E1001 09:38:14.523387 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:38:14 crc kubenswrapper[4787]: E1001 09:38:14.523545 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:38:15 crc kubenswrapper[4787]: I1001 09:38:15.522791 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:38:15 crc kubenswrapper[4787]: E1001 09:38:15.523499 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:38:16 crc kubenswrapper[4787]: I1001 09:38:16.523180 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:38:16 crc kubenswrapper[4787]: I1001 09:38:16.523266 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:38:16 crc kubenswrapper[4787]: I1001 09:38:16.523320 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:38:16 crc kubenswrapper[4787]: E1001 09:38:16.523336 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:38:16 crc kubenswrapper[4787]: E1001 09:38:16.523466 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:38:16 crc kubenswrapper[4787]: E1001 09:38:16.523604 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:38:17 crc kubenswrapper[4787]: I1001 09:38:17.523556 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:38:17 crc kubenswrapper[4787]: E1001 09:38:17.523797 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:38:17 crc kubenswrapper[4787]: I1001 09:38:17.523929 4787 scope.go:117] "RemoveContainer" containerID="8c830596f83deb4fa998fa9af881c7eb1f3673588f444edc9cbf895c19b8248c" Oct 01 09:38:18 crc kubenswrapper[4787]: I1001 09:38:18.167534 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z8tjz_53f5ffd2-7649-4c9e-929f-a68187409644/kube-multus/1.log" Oct 01 09:38:18 crc kubenswrapper[4787]: I1001 09:38:18.167638 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-z8tjz" event={"ID":"53f5ffd2-7649-4c9e-929f-a68187409644","Type":"ContainerStarted","Data":"8ccabd0db928904435aacf77b37e997bf489a08e4cc113b3438d986e0c7f0ed2"} Oct 01 09:38:18 crc kubenswrapper[4787]: I1001 09:38:18.523492 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:38:18 crc kubenswrapper[4787]: I1001 09:38:18.523642 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:38:18 crc kubenswrapper[4787]: I1001 09:38:18.525219 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:38:18 crc kubenswrapper[4787]: E1001 09:38:18.525386 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:38:18 crc kubenswrapper[4787]: E1001 09:38:18.525601 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:38:18 crc kubenswrapper[4787]: E1001 09:38:18.525218 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:38:18 crc kubenswrapper[4787]: E1001 09:38:18.638560 4787 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 01 09:38:19 crc kubenswrapper[4787]: I1001 09:38:19.522774 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:38:19 crc kubenswrapper[4787]: E1001 09:38:19.523273 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:38:20 crc kubenswrapper[4787]: I1001 09:38:20.523386 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:38:20 crc kubenswrapper[4787]: I1001 09:38:20.523452 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:38:20 crc kubenswrapper[4787]: I1001 09:38:20.523452 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:38:20 crc kubenswrapper[4787]: E1001 09:38:20.523571 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:38:20 crc kubenswrapper[4787]: E1001 09:38:20.523731 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:38:20 crc kubenswrapper[4787]: E1001 09:38:20.523861 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:38:21 crc kubenswrapper[4787]: I1001 09:38:21.522951 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:38:21 crc kubenswrapper[4787]: E1001 09:38:21.523105 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:38:22 crc kubenswrapper[4787]: I1001 09:38:22.523308 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:38:22 crc kubenswrapper[4787]: E1001 09:38:22.523432 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6zz6g" podUID="0eb4a902-9efa-4088-a236-9a634fd085fd" Oct 01 09:38:22 crc kubenswrapper[4787]: I1001 09:38:22.523309 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:38:22 crc kubenswrapper[4787]: I1001 09:38:22.523606 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:38:22 crc kubenswrapper[4787]: E1001 09:38:22.523662 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 01 09:38:22 crc kubenswrapper[4787]: E1001 09:38:22.523600 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 01 09:38:23 crc kubenswrapper[4787]: I1001 09:38:23.523807 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:38:23 crc kubenswrapper[4787]: E1001 09:38:23.524333 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 01 09:38:24 crc kubenswrapper[4787]: I1001 09:38:24.523447 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:38:24 crc kubenswrapper[4787]: I1001 09:38:24.523545 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:38:24 crc kubenswrapper[4787]: I1001 09:38:24.523677 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:38:24 crc kubenswrapper[4787]: I1001 09:38:24.528846 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 01 09:38:24 crc kubenswrapper[4787]: I1001 09:38:24.530955 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 01 09:38:24 crc kubenswrapper[4787]: I1001 09:38:24.532662 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 01 09:38:24 crc kubenswrapper[4787]: I1001 09:38:24.533126 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 01 09:38:24 crc kubenswrapper[4787]: I1001 09:38:24.533502 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 01 09:38:24 crc kubenswrapper[4787]: I1001 09:38:24.534159 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 01 09:38:25 crc kubenswrapper[4787]: I1001 09:38:25.523169 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.005421 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.048323 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-djtn2"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.048896 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-djtn2" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.051471 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-4g5xc"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.052078 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-4g5xc" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.053221 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhp4"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.053537 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhp4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.055847 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-x649j"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.056417 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-x649j" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.056491 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-cdc2f"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.057128 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-cdc2f" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.066313 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.072622 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tvvw6"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.073226 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tvvw6" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.076947 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.078158 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.078650 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.079169 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.079527 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.079863 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-tc862"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.080308 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-hwftw"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.080611 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.080653 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.080861 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.081017 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jczvb"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.081149 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.081167 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-tc862" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.081266 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.081745 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.081914 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.082989 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.083987 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.084682 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.085493 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.085936 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.087162 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.087593 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.087671 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.087695 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.087617 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.087934 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.088288 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.088319 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.109938 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jczvb" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.110488 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.110676 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.110836 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.111012 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.111558 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.110484 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cpqnn"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.111573 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.111872 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.111958 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.112029 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.112188 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.112277 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.113125 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.113290 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.113301 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.114975 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.116076 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.116832 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.116947 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.117798 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.118651 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.119438 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-7p4xb"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.119988 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-7p4xb" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.120691 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-gjtg8"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.121239 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gjtg8" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.123161 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dscrd"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.123722 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dscrd" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.124767 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.126487 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.142823 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-tzsx8"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.143489 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-t2mbp"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.143962 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t2mbp" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.144067 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-46vzw"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.144418 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tzsx8" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.144691 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-46vzw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.151119 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.152252 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fzd5j"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.153143 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fzd5j" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.164163 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/07892f22-556b-49a6-8f71-3d3b16b2cdef-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-4g5xc\" (UID: \"07892f22-556b-49a6-8f71-3d3b16b2cdef\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4g5xc" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.164209 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e6f1b5a2-581a-4c1a-b952-5662b2fb636f-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-x649j\" (UID: \"e6f1b5a2-581a-4c1a-b952-5662b2fb636f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-x649j" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.164243 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-628nj\" (UniqueName: \"kubernetes.io/projected/e6f1b5a2-581a-4c1a-b952-5662b2fb636f-kube-api-access-628nj\") pod \"controller-manager-879f6c89f-x649j\" (UID: \"e6f1b5a2-581a-4c1a-b952-5662b2fb636f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-x649j" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.164267 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/160b7ce2-3861-4c15-af76-2b0d41ec8301-audit-dir\") pod \"apiserver-7bbb656c7d-hlhp4\" (UID: \"160b7ce2-3861-4c15-af76-2b0d41ec8301\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhp4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.164292 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6f1b5a2-581a-4c1a-b952-5662b2fb636f-config\") pod \"controller-manager-879f6c89f-x649j\" (UID: \"e6f1b5a2-581a-4c1a-b952-5662b2fb636f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-x649j" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.164315 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6f1b5a2-581a-4c1a-b952-5662b2fb636f-serving-cert\") pod \"controller-manager-879f6c89f-x649j\" (UID: \"e6f1b5a2-581a-4c1a-b952-5662b2fb636f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-x649j" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.164337 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/160b7ce2-3861-4c15-af76-2b0d41ec8301-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-hlhp4\" (UID: \"160b7ce2-3861-4c15-af76-2b0d41ec8301\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhp4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.164359 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frkgv\" (UniqueName: \"kubernetes.io/projected/160b7ce2-3861-4c15-af76-2b0d41ec8301-kube-api-access-frkgv\") pod \"apiserver-7bbb656c7d-hlhp4\" (UID: \"160b7ce2-3861-4c15-af76-2b0d41ec8301\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhp4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.164382 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2pgd\" (UniqueName: \"kubernetes.io/projected/07892f22-556b-49a6-8f71-3d3b16b2cdef-kube-api-access-t2pgd\") pod \"machine-api-operator-5694c8668f-4g5xc\" (UID: \"07892f22-556b-49a6-8f71-3d3b16b2cdef\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4g5xc" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.164408 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5f699a8f-ffe8-4512-9e55-24c88e4154d2-client-ca\") pod \"route-controller-manager-6576b87f9c-djtn2\" (UID: \"5f699a8f-ffe8-4512-9e55-24c88e4154d2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-djtn2" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.164431 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89wcv\" (UniqueName: \"kubernetes.io/projected/5f699a8f-ffe8-4512-9e55-24c88e4154d2-kube-api-access-89wcv\") pod \"route-controller-manager-6576b87f9c-djtn2\" (UID: \"5f699a8f-ffe8-4512-9e55-24c88e4154d2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-djtn2" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.164466 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fad32d96-ad78-48b4-9874-bef9a0a30f8f-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-cdc2f\" (UID: \"fad32d96-ad78-48b4-9874-bef9a0a30f8f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cdc2f" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.164511 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/07892f22-556b-49a6-8f71-3d3b16b2cdef-images\") pod \"machine-api-operator-5694c8668f-4g5xc\" (UID: \"07892f22-556b-49a6-8f71-3d3b16b2cdef\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4g5xc" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.164533 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbcfn\" (UniqueName: \"kubernetes.io/projected/fad32d96-ad78-48b4-9874-bef9a0a30f8f-kube-api-access-xbcfn\") pod \"authentication-operator-69f744f599-cdc2f\" (UID: \"fad32d96-ad78-48b4-9874-bef9a0a30f8f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cdc2f" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.164594 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/160b7ce2-3861-4c15-af76-2b0d41ec8301-audit-policies\") pod \"apiserver-7bbb656c7d-hlhp4\" (UID: \"160b7ce2-3861-4c15-af76-2b0d41ec8301\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhp4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.164619 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f699a8f-ffe8-4512-9e55-24c88e4154d2-config\") pod \"route-controller-manager-6576b87f9c-djtn2\" (UID: \"5f699a8f-ffe8-4512-9e55-24c88e4154d2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-djtn2" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.164642 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fad32d96-ad78-48b4-9874-bef9a0a30f8f-service-ca-bundle\") pod \"authentication-operator-69f744f599-cdc2f\" (UID: \"fad32d96-ad78-48b4-9874-bef9a0a30f8f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cdc2f" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.164662 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fad32d96-ad78-48b4-9874-bef9a0a30f8f-config\") pod \"authentication-operator-69f744f599-cdc2f\" (UID: \"fad32d96-ad78-48b4-9874-bef9a0a30f8f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cdc2f" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.164702 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/160b7ce2-3861-4c15-af76-2b0d41ec8301-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-hlhp4\" (UID: \"160b7ce2-3861-4c15-af76-2b0d41ec8301\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhp4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.164723 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07892f22-556b-49a6-8f71-3d3b16b2cdef-config\") pod \"machine-api-operator-5694c8668f-4g5xc\" (UID: \"07892f22-556b-49a6-8f71-3d3b16b2cdef\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4g5xc" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.164748 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fad32d96-ad78-48b4-9874-bef9a0a30f8f-serving-cert\") pod \"authentication-operator-69f744f599-cdc2f\" (UID: \"fad32d96-ad78-48b4-9874-bef9a0a30f8f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cdc2f" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.164767 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/160b7ce2-3861-4c15-af76-2b0d41ec8301-serving-cert\") pod \"apiserver-7bbb656c7d-hlhp4\" (UID: \"160b7ce2-3861-4c15-af76-2b0d41ec8301\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhp4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.164791 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/160b7ce2-3861-4c15-af76-2b0d41ec8301-etcd-client\") pod \"apiserver-7bbb656c7d-hlhp4\" (UID: \"160b7ce2-3861-4c15-af76-2b0d41ec8301\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhp4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.164812 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/160b7ce2-3861-4c15-af76-2b0d41ec8301-encryption-config\") pod \"apiserver-7bbb656c7d-hlhp4\" (UID: \"160b7ce2-3861-4c15-af76-2b0d41ec8301\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhp4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.164836 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e6f1b5a2-581a-4c1a-b952-5662b2fb636f-client-ca\") pod \"controller-manager-879f6c89f-x649j\" (UID: \"e6f1b5a2-581a-4c1a-b952-5662b2fb636f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-x649j" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.164857 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5f699a8f-ffe8-4512-9e55-24c88e4154d2-serving-cert\") pod \"route-controller-manager-6576b87f9c-djtn2\" (UID: \"5f699a8f-ffe8-4512-9e55-24c88e4154d2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-djtn2" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.171223 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-lld4x"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.171373 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.171931 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.172155 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-pzdzn"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.191103 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pzdzn" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.195303 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-lld4x" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.198562 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.198621 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qvdw4"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.204123 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.205682 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.205732 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.205909 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.222460 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.222793 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.224172 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-hqx7l"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.224536 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-hqx7l" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.225382 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.225789 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qvdw4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.225806 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.225839 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.225861 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.226015 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.226066 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.226742 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.227320 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.227423 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.227616 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.227909 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.228010 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.228168 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.228399 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.230215 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.230421 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.230698 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.231201 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.232406 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.232623 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.233046 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.233209 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.233362 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.233486 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.233605 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.233787 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.234000 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.234103 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.234157 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.234272 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.234409 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.234519 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.234531 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.235990 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.236152 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.236249 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.236443 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.238475 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.238614 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.238743 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.238855 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.238963 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.239307 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-fsblz"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.240315 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-n47zx"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.240724 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9x4nl"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.241133 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9x4nl" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.241328 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-fsblz" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.241503 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n47zx" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.243921 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.246949 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.247415 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.247629 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.248053 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.248968 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.252430 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.253134 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-29mhg"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.254098 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-29mhg" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.255503 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.263896 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.266827 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/07892f22-556b-49a6-8f71-3d3b16b2cdef-images\") pod \"machine-api-operator-5694c8668f-4g5xc\" (UID: \"07892f22-556b-49a6-8f71-3d3b16b2cdef\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4g5xc" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.266875 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbcfn\" (UniqueName: \"kubernetes.io/projected/fad32d96-ad78-48b4-9874-bef9a0a30f8f-kube-api-access-xbcfn\") pod \"authentication-operator-69f744f599-cdc2f\" (UID: \"fad32d96-ad78-48b4-9874-bef9a0a30f8f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cdc2f" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.266902 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/160b7ce2-3861-4c15-af76-2b0d41ec8301-audit-policies\") pod \"apiserver-7bbb656c7d-hlhp4\" (UID: \"160b7ce2-3861-4c15-af76-2b0d41ec8301\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhp4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.266939 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f699a8f-ffe8-4512-9e55-24c88e4154d2-config\") pod \"route-controller-manager-6576b87f9c-djtn2\" (UID: \"5f699a8f-ffe8-4512-9e55-24c88e4154d2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-djtn2" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.266965 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fad32d96-ad78-48b4-9874-bef9a0a30f8f-service-ca-bundle\") pod \"authentication-operator-69f744f599-cdc2f\" (UID: \"fad32d96-ad78-48b4-9874-bef9a0a30f8f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cdc2f" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.266985 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fad32d96-ad78-48b4-9874-bef9a0a30f8f-config\") pod \"authentication-operator-69f744f599-cdc2f\" (UID: \"fad32d96-ad78-48b4-9874-bef9a0a30f8f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cdc2f" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.267012 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzwr2\" (UniqueName: \"kubernetes.io/projected/52a18549-2c1c-4f28-bfa1-6fb14e0690e1-kube-api-access-hzwr2\") pod \"router-default-5444994796-7p4xb\" (UID: \"52a18549-2c1c-4f28-bfa1-6fb14e0690e1\") " pod="openshift-ingress/router-default-5444994796-7p4xb" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.267054 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/160b7ce2-3861-4c15-af76-2b0d41ec8301-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-hlhp4\" (UID: \"160b7ce2-3861-4c15-af76-2b0d41ec8301\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhp4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.267097 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/52a18549-2c1c-4f28-bfa1-6fb14e0690e1-metrics-certs\") pod \"router-default-5444994796-7p4xb\" (UID: \"52a18549-2c1c-4f28-bfa1-6fb14e0690e1\") " pod="openshift-ingress/router-default-5444994796-7p4xb" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.267123 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07892f22-556b-49a6-8f71-3d3b16b2cdef-config\") pod \"machine-api-operator-5694c8668f-4g5xc\" (UID: \"07892f22-556b-49a6-8f71-3d3b16b2cdef\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4g5xc" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.267147 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k84xq\" (UniqueName: \"kubernetes.io/projected/15f1cb19-94ee-49b2-a837-b7bdaf484387-kube-api-access-k84xq\") pod \"migrator-59844c95c7-t2mbp\" (UID: \"15f1cb19-94ee-49b2-a837-b7bdaf484387\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t2mbp" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.267174 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fad32d96-ad78-48b4-9874-bef9a0a30f8f-serving-cert\") pod \"authentication-operator-69f744f599-cdc2f\" (UID: \"fad32d96-ad78-48b4-9874-bef9a0a30f8f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cdc2f" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.267199 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/160b7ce2-3861-4c15-af76-2b0d41ec8301-serving-cert\") pod \"apiserver-7bbb656c7d-hlhp4\" (UID: \"160b7ce2-3861-4c15-af76-2b0d41ec8301\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhp4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.267231 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/52a18549-2c1c-4f28-bfa1-6fb14e0690e1-default-certificate\") pod \"router-default-5444994796-7p4xb\" (UID: \"52a18549-2c1c-4f28-bfa1-6fb14e0690e1\") " pod="openshift-ingress/router-default-5444994796-7p4xb" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.267255 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/160b7ce2-3861-4c15-af76-2b0d41ec8301-etcd-client\") pod \"apiserver-7bbb656c7d-hlhp4\" (UID: \"160b7ce2-3861-4c15-af76-2b0d41ec8301\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhp4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.267285 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/52a18549-2c1c-4f28-bfa1-6fb14e0690e1-stats-auth\") pod \"router-default-5444994796-7p4xb\" (UID: \"52a18549-2c1c-4f28-bfa1-6fb14e0690e1\") " pod="openshift-ingress/router-default-5444994796-7p4xb" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.267314 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/160b7ce2-3861-4c15-af76-2b0d41ec8301-encryption-config\") pod \"apiserver-7bbb656c7d-hlhp4\" (UID: \"160b7ce2-3861-4c15-af76-2b0d41ec8301\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhp4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.267333 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rv6xx"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.267958 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/160b7ce2-3861-4c15-af76-2b0d41ec8301-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-hlhp4\" (UID: \"160b7ce2-3861-4c15-af76-2b0d41ec8301\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhp4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.268058 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/07892f22-556b-49a6-8f71-3d3b16b2cdef-images\") pod \"machine-api-operator-5694c8668f-4g5xc\" (UID: \"07892f22-556b-49a6-8f71-3d3b16b2cdef\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4g5xc" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.268159 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rv6xx" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.268293 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e6f1b5a2-581a-4c1a-b952-5662b2fb636f-client-ca\") pod \"controller-manager-879f6c89f-x649j\" (UID: \"e6f1b5a2-581a-4c1a-b952-5662b2fb636f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-x649j" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.267344 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e6f1b5a2-581a-4c1a-b952-5662b2fb636f-client-ca\") pod \"controller-manager-879f6c89f-x649j\" (UID: \"e6f1b5a2-581a-4c1a-b952-5662b2fb636f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-x649j" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.268588 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5f699a8f-ffe8-4512-9e55-24c88e4154d2-serving-cert\") pod \"route-controller-manager-6576b87f9c-djtn2\" (UID: \"5f699a8f-ffe8-4512-9e55-24c88e4154d2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-djtn2" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.268634 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/52a18549-2c1c-4f28-bfa1-6fb14e0690e1-service-ca-bundle\") pod \"router-default-5444994796-7p4xb\" (UID: \"52a18549-2c1c-4f28-bfa1-6fb14e0690e1\") " pod="openshift-ingress/router-default-5444994796-7p4xb" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.268676 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/07892f22-556b-49a6-8f71-3d3b16b2cdef-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-4g5xc\" (UID: \"07892f22-556b-49a6-8f71-3d3b16b2cdef\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4g5xc" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.268681 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07892f22-556b-49a6-8f71-3d3b16b2cdef-config\") pod \"machine-api-operator-5694c8668f-4g5xc\" (UID: \"07892f22-556b-49a6-8f71-3d3b16b2cdef\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4g5xc" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.268706 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e6f1b5a2-581a-4c1a-b952-5662b2fb636f-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-x649j\" (UID: \"e6f1b5a2-581a-4c1a-b952-5662b2fb636f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-x649j" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.268735 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-628nj\" (UniqueName: \"kubernetes.io/projected/e6f1b5a2-581a-4c1a-b952-5662b2fb636f-kube-api-access-628nj\") pod \"controller-manager-879f6c89f-x649j\" (UID: \"e6f1b5a2-581a-4c1a-b952-5662b2fb636f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-x649j" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.268761 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/160b7ce2-3861-4c15-af76-2b0d41ec8301-audit-dir\") pod \"apiserver-7bbb656c7d-hlhp4\" (UID: \"160b7ce2-3861-4c15-af76-2b0d41ec8301\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhp4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.268792 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6f1b5a2-581a-4c1a-b952-5662b2fb636f-config\") pod \"controller-manager-879f6c89f-x649j\" (UID: \"e6f1b5a2-581a-4c1a-b952-5662b2fb636f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-x649j" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.268817 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/1c07acd7-0397-494c-ab52-83bcdefdf2ac-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-tvvw6\" (UID: \"1c07acd7-0397-494c-ab52-83bcdefdf2ac\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tvvw6" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.268852 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6f1b5a2-581a-4c1a-b952-5662b2fb636f-serving-cert\") pod \"controller-manager-879f6c89f-x649j\" (UID: \"e6f1b5a2-581a-4c1a-b952-5662b2fb636f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-x649j" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.268878 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/160b7ce2-3861-4c15-af76-2b0d41ec8301-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-hlhp4\" (UID: \"160b7ce2-3861-4c15-af76-2b0d41ec8301\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhp4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.268900 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frkgv\" (UniqueName: \"kubernetes.io/projected/160b7ce2-3861-4c15-af76-2b0d41ec8301-kube-api-access-frkgv\") pod \"apiserver-7bbb656c7d-hlhp4\" (UID: \"160b7ce2-3861-4c15-af76-2b0d41ec8301\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhp4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.268925 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2pgd\" (UniqueName: \"kubernetes.io/projected/07892f22-556b-49a6-8f71-3d3b16b2cdef-kube-api-access-t2pgd\") pod \"machine-api-operator-5694c8668f-4g5xc\" (UID: \"07892f22-556b-49a6-8f71-3d3b16b2cdef\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4g5xc" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.268952 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmwmk\" (UniqueName: \"kubernetes.io/projected/1c07acd7-0397-494c-ab52-83bcdefdf2ac-kube-api-access-mmwmk\") pod \"cluster-samples-operator-665b6dd947-tvvw6\" (UID: \"1c07acd7-0397-494c-ab52-83bcdefdf2ac\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tvvw6" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.268981 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5f699a8f-ffe8-4512-9e55-24c88e4154d2-client-ca\") pod \"route-controller-manager-6576b87f9c-djtn2\" (UID: \"5f699a8f-ffe8-4512-9e55-24c88e4154d2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-djtn2" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.269006 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89wcv\" (UniqueName: \"kubernetes.io/projected/5f699a8f-ffe8-4512-9e55-24c88e4154d2-kube-api-access-89wcv\") pod \"route-controller-manager-6576b87f9c-djtn2\" (UID: \"5f699a8f-ffe8-4512-9e55-24c88e4154d2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-djtn2" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.269053 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fad32d96-ad78-48b4-9874-bef9a0a30f8f-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-cdc2f\" (UID: \"fad32d96-ad78-48b4-9874-bef9a0a30f8f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cdc2f" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.269426 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fad32d96-ad78-48b4-9874-bef9a0a30f8f-service-ca-bundle\") pod \"authentication-operator-69f744f599-cdc2f\" (UID: \"fad32d96-ad78-48b4-9874-bef9a0a30f8f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cdc2f" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.269496 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fad32d96-ad78-48b4-9874-bef9a0a30f8f-config\") pod \"authentication-operator-69f744f599-cdc2f\" (UID: \"fad32d96-ad78-48b4-9874-bef9a0a30f8f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cdc2f" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.269692 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f699a8f-ffe8-4512-9e55-24c88e4154d2-config\") pod \"route-controller-manager-6576b87f9c-djtn2\" (UID: \"5f699a8f-ffe8-4512-9e55-24c88e4154d2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-djtn2" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.270659 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/160b7ce2-3861-4c15-af76-2b0d41ec8301-audit-dir\") pod \"apiserver-7bbb656c7d-hlhp4\" (UID: \"160b7ce2-3861-4c15-af76-2b0d41ec8301\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhp4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.270815 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5f699a8f-ffe8-4512-9e55-24c88e4154d2-client-ca\") pod \"route-controller-manager-6576b87f9c-djtn2\" (UID: \"5f699a8f-ffe8-4512-9e55-24c88e4154d2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-djtn2" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.268760 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-86dvz"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.271505 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-whc4s"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.271956 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-whc4s" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.272056 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-86dvz" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.272253 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/160b7ce2-3861-4c15-af76-2b0d41ec8301-audit-policies\") pod \"apiserver-7bbb656c7d-hlhp4\" (UID: \"160b7ce2-3861-4c15-af76-2b0d41ec8301\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhp4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.272430 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r8xzr"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.272672 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6f1b5a2-581a-4c1a-b952-5662b2fb636f-config\") pod \"controller-manager-879f6c89f-x649j\" (UID: \"e6f1b5a2-581a-4c1a-b952-5662b2fb636f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-x649j" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.272827 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/160b7ce2-3861-4c15-af76-2b0d41ec8301-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-hlhp4\" (UID: \"160b7ce2-3861-4c15-af76-2b0d41ec8301\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhp4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.273111 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r8xzr" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.277189 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e6f1b5a2-581a-4c1a-b952-5662b2fb636f-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-x649j\" (UID: \"e6f1b5a2-581a-4c1a-b952-5662b2fb636f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-x649j" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.278919 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-n6959"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.279738 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-n6959" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.279772 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6f1b5a2-581a-4c1a-b952-5662b2fb636f-serving-cert\") pod \"controller-manager-879f6c89f-x649j\" (UID: \"e6f1b5a2-581a-4c1a-b952-5662b2fb636f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-x649j" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.279794 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fad32d96-ad78-48b4-9874-bef9a0a30f8f-serving-cert\") pod \"authentication-operator-69f744f599-cdc2f\" (UID: \"fad32d96-ad78-48b4-9874-bef9a0a30f8f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cdc2f" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.279825 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/160b7ce2-3861-4c15-af76-2b0d41ec8301-serving-cert\") pod \"apiserver-7bbb656c7d-hlhp4\" (UID: \"160b7ce2-3861-4c15-af76-2b0d41ec8301\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhp4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.280051 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.281175 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zssrz"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.281670 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/07892f22-556b-49a6-8f71-3d3b16b2cdef-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-4g5xc\" (UID: \"07892f22-556b-49a6-8f71-3d3b16b2cdef\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4g5xc" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.281959 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zssrz" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.282244 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321850-kw99x"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.282722 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-kw99x" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.283147 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fad32d96-ad78-48b4-9874-bef9a0a30f8f-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-cdc2f\" (UID: \"fad32d96-ad78-48b4-9874-bef9a0a30f8f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cdc2f" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.283816 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-4g5xc"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.307158 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/160b7ce2-3861-4c15-af76-2b0d41ec8301-encryption-config\") pod \"apiserver-7bbb656c7d-hlhp4\" (UID: \"160b7ce2-3861-4c15-af76-2b0d41ec8301\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhp4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.307618 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5f699a8f-ffe8-4512-9e55-24c88e4154d2-serving-cert\") pod \"route-controller-manager-6576b87f9c-djtn2\" (UID: \"5f699a8f-ffe8-4512-9e55-24c88e4154d2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-djtn2" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.309659 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-tn48z"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.310193 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/160b7ce2-3861-4c15-af76-2b0d41ec8301-etcd-client\") pod \"apiserver-7bbb656c7d-hlhp4\" (UID: \"160b7ce2-3861-4c15-af76-2b0d41ec8301\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhp4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.310757 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-djtn2"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.310882 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-tn48z" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.312346 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.312417 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-gqvt9"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.313313 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-gqvt9" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.314010 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-blwqt"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.315182 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-sgphp"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.315583 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-blwqt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.318353 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-sgphp" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.326577 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-hqjh4"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.327590 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hqjh4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.328200 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-x649j"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.331218 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.331442 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cf76m"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.332908 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cf76m" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.339249 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhp4"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.343336 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tvvw6"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.345046 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-tc862"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.347167 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-cdc2f"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.347427 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jczvb"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.348638 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-hwftw"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.349492 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-ns9tt"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.350296 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-ns9tt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.350983 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.351939 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-tzsx8"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.352752 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-lld4x"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.353903 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-n6959"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.355574 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-hqjh4"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.356211 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-gqvt9"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.356947 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-fsblz"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.357860 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-tn48z"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.358841 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-t2mbp"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.360155 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-86dvz"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.360778 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cpqnn"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.361774 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zssrz"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.362736 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-gjtg8"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.363825 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-n47zx"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.364833 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dscrd"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.366903 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fzd5j"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.368042 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-46vzw"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.369240 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-sgxp6"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.369952 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzwr2\" (UniqueName: \"kubernetes.io/projected/52a18549-2c1c-4f28-bfa1-6fb14e0690e1-kube-api-access-hzwr2\") pod \"router-default-5444994796-7p4xb\" (UID: \"52a18549-2c1c-4f28-bfa1-6fb14e0690e1\") " pod="openshift-ingress/router-default-5444994796-7p4xb" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.370035 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/52a18549-2c1c-4f28-bfa1-6fb14e0690e1-metrics-certs\") pod \"router-default-5444994796-7p4xb\" (UID: \"52a18549-2c1c-4f28-bfa1-6fb14e0690e1\") " pod="openshift-ingress/router-default-5444994796-7p4xb" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.370070 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k84xq\" (UniqueName: \"kubernetes.io/projected/15f1cb19-94ee-49b2-a837-b7bdaf484387-kube-api-access-k84xq\") pod \"migrator-59844c95c7-t2mbp\" (UID: \"15f1cb19-94ee-49b2-a837-b7bdaf484387\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t2mbp" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.370121 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/52a18549-2c1c-4f28-bfa1-6fb14e0690e1-default-certificate\") pod \"router-default-5444994796-7p4xb\" (UID: \"52a18549-2c1c-4f28-bfa1-6fb14e0690e1\") " pod="openshift-ingress/router-default-5444994796-7p4xb" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.370137 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/52a18549-2c1c-4f28-bfa1-6fb14e0690e1-stats-auth\") pod \"router-default-5444994796-7p4xb\" (UID: \"52a18549-2c1c-4f28-bfa1-6fb14e0690e1\") " pod="openshift-ingress/router-default-5444994796-7p4xb" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.370159 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/52a18549-2c1c-4f28-bfa1-6fb14e0690e1-service-ca-bundle\") pod \"router-default-5444994796-7p4xb\" (UID: \"52a18549-2c1c-4f28-bfa1-6fb14e0690e1\") " pod="openshift-ingress/router-default-5444994796-7p4xb" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.370184 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/1c07acd7-0397-494c-ab52-83bcdefdf2ac-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-tvvw6\" (UID: \"1c07acd7-0397-494c-ab52-83bcdefdf2ac\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tvvw6" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.370215 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmwmk\" (UniqueName: \"kubernetes.io/projected/1c07acd7-0397-494c-ab52-83bcdefdf2ac-kube-api-access-mmwmk\") pod \"cluster-samples-operator-665b6dd947-tvvw6\" (UID: \"1c07acd7-0397-494c-ab52-83bcdefdf2ac\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tvvw6" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.370599 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-sgxp6" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.371260 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-9lvh8"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.371413 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.371636 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/52a18549-2c1c-4f28-bfa1-6fb14e0690e1-service-ca-bundle\") pod \"router-default-5444994796-7p4xb\" (UID: \"52a18549-2c1c-4f28-bfa1-6fb14e0690e1\") " pod="openshift-ingress/router-default-5444994796-7p4xb" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.372104 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-9lvh8" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.372674 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r8xzr"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.373880 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-hqx7l"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.373985 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/52a18549-2c1c-4f28-bfa1-6fb14e0690e1-stats-auth\") pod \"router-default-5444994796-7p4xb\" (UID: \"52a18549-2c1c-4f28-bfa1-6fb14e0690e1\") " pod="openshift-ingress/router-default-5444994796-7p4xb" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.374640 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/1c07acd7-0397-494c-ab52-83bcdefdf2ac-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-tvvw6\" (UID: \"1c07acd7-0397-494c-ab52-83bcdefdf2ac\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tvvw6" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.375654 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qvdw4"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.376603 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/52a18549-2c1c-4f28-bfa1-6fb14e0690e1-default-certificate\") pod \"router-default-5444994796-7p4xb\" (UID: \"52a18549-2c1c-4f28-bfa1-6fb14e0690e1\") " pod="openshift-ingress/router-default-5444994796-7p4xb" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.376811 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9x4nl"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.377836 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/52a18549-2c1c-4f28-bfa1-6fb14e0690e1-metrics-certs\") pod \"router-default-5444994796-7p4xb\" (UID: \"52a18549-2c1c-4f28-bfa1-6fb14e0690e1\") " pod="openshift-ingress/router-default-5444994796-7p4xb" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.378014 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rv6xx"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.379029 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-sgxp6"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.380056 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-29mhg"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.381242 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321850-kw99x"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.382248 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-sgphp"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.383357 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-whc4s"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.384382 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-blwqt"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.385429 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cf76m"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.386470 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-9lvh8"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.387569 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-ncnkh"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.390438 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-ncnkh"] Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.390680 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-ncnkh" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.392186 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.421738 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.431749 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.451307 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.470797 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.491236 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.511043 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.532589 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.571414 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.571756 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.571824 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.571887 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.571923 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.571953 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/82a8aff4-187a-464a-b375-5c20e576164d-proxy-tls\") pod \"machine-config-operator-74547568cd-gjtg8\" (UID: \"82a8aff4-187a-464a-b375-5c20e576164d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gjtg8" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.571977 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgmrq\" (UniqueName: \"kubernetes.io/projected/8ffe49e7-9382-4711-985c-8d6f76c42abc-kube-api-access-dgmrq\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.572012 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/b0ea3010-e975-4d1e-9fa4-385489660daa-image-import-ca\") pod \"apiserver-76f77b778f-46vzw\" (UID: \"b0ea3010-e975-4d1e-9fa4-385489660daa\") " pod="openshift-apiserver/apiserver-76f77b778f-46vzw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.572042 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-audit-dir\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.572105 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/aa998110-e80e-4bc1-b1b2-f9806030ba26-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-jczvb\" (UID: \"aa998110-e80e-4bc1-b1b2-f9806030ba26\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jczvb" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.572128 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b0ea3010-e975-4d1e-9fa4-385489660daa-node-pullsecrets\") pod \"apiserver-76f77b778f-46vzw\" (UID: \"b0ea3010-e975-4d1e-9fa4-385489660daa\") " pod="openshift-apiserver/apiserver-76f77b778f-46vzw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.572170 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8ffe49e7-9382-4711-985c-8d6f76c42abc-ca-trust-extracted\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.572219 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/57a3e335-0a9a-4fb5-b234-e0017b775ecd-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-dscrd\" (UID: \"57a3e335-0a9a-4fb5-b234-e0017b775ecd\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dscrd" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.572296 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bvqz\" (UniqueName: \"kubernetes.io/projected/fa4ba99a-cfa6-4e3b-bd6f-ce18ea3fefe8-kube-api-access-6bvqz\") pod \"machine-config-controller-84d6567774-tzsx8\" (UID: \"fa4ba99a-cfa6-4e3b-bd6f-ce18ea3fefe8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tzsx8" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.572358 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aa998110-e80e-4bc1-b1b2-f9806030ba26-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-jczvb\" (UID: \"aa998110-e80e-4bc1-b1b2-f9806030ba26\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jczvb" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.572387 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57a3e335-0a9a-4fb5-b234-e0017b775ecd-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-dscrd\" (UID: \"57a3e335-0a9a-4fb5-b234-e0017b775ecd\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dscrd" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.572440 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.572595 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b0ea3010-e975-4d1e-9fa4-385489660daa-encryption-config\") pod \"apiserver-76f77b778f-46vzw\" (UID: \"b0ea3010-e975-4d1e-9fa4-385489660daa\") " pod="openshift-apiserver/apiserver-76f77b778f-46vzw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.572687 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srs6n\" (UniqueName: \"kubernetes.io/projected/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-kube-api-access-srs6n\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.572784 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhq4f\" (UniqueName: \"kubernetes.io/projected/652585ff-b553-4326-add1-ba20d0dd0a11-kube-api-access-dhq4f\") pod \"openshift-controller-manager-operator-756b6f6bc6-fzd5j\" (UID: \"652585ff-b553-4326-add1-ba20d0dd0a11\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fzd5j" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.572871 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/652585ff-b553-4326-add1-ba20d0dd0a11-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-fzd5j\" (UID: \"652585ff-b553-4326-add1-ba20d0dd0a11\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fzd5j" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.572963 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b0ea3010-e975-4d1e-9fa4-385489660daa-etcd-client\") pod \"apiserver-76f77b778f-46vzw\" (UID: \"b0ea3010-e975-4d1e-9fa4-385489660daa\") " pod="openshift-apiserver/apiserver-76f77b778f-46vzw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.573050 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.573138 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.573163 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86g4f\" (UniqueName: \"kubernetes.io/projected/b0ea3010-e975-4d1e-9fa4-385489660daa-kube-api-access-86g4f\") pod \"apiserver-76f77b778f-46vzw\" (UID: \"b0ea3010-e975-4d1e-9fa4-385489660daa\") " pod="openshift-apiserver/apiserver-76f77b778f-46vzw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.573182 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.573218 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlccc\" (UniqueName: \"kubernetes.io/projected/aa998110-e80e-4bc1-b1b2-f9806030ba26-kube-api-access-vlccc\") pod \"cluster-image-registry-operator-dc59b4c8b-jczvb\" (UID: \"aa998110-e80e-4bc1-b1b2-f9806030ba26\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jczvb" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.573239 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.573316 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fa4ba99a-cfa6-4e3b-bd6f-ce18ea3fefe8-proxy-tls\") pod \"machine-config-controller-84d6567774-tzsx8\" (UID: \"fa4ba99a-cfa6-4e3b-bd6f-ce18ea3fefe8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tzsx8" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.573362 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/b0ea3010-e975-4d1e-9fa4-385489660daa-audit\") pod \"apiserver-76f77b778f-46vzw\" (UID: \"b0ea3010-e975-4d1e-9fa4-385489660daa\") " pod="openshift-apiserver/apiserver-76f77b778f-46vzw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.573392 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjbhw\" (UniqueName: \"kubernetes.io/projected/82a8aff4-187a-464a-b375-5c20e576164d-kube-api-access-hjbhw\") pod \"machine-config-operator-74547568cd-gjtg8\" (UID: \"82a8aff4-187a-464a-b375-5c20e576164d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gjtg8" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.573447 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8ffe49e7-9382-4711-985c-8d6f76c42abc-bound-sa-token\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.573478 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/82a8aff4-187a-464a-b375-5c20e576164d-images\") pod \"machine-config-operator-74547568cd-gjtg8\" (UID: \"82a8aff4-187a-464a-b375-5c20e576164d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gjtg8" Oct 01 09:38:27 crc kubenswrapper[4787]: E1001 09:38:27.573649 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:28.073624714 +0000 UTC m=+140.188768911 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.573777 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8ffe49e7-9382-4711-985c-8d6f76c42abc-registry-tls\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.573834 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fa4ba99a-cfa6-4e3b-bd6f-ce18ea3fefe8-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-tzsx8\" (UID: \"fa4ba99a-cfa6-4e3b-bd6f-ce18ea3fefe8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tzsx8" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.573908 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b0ea3010-e975-4d1e-9fa4-385489660daa-serving-cert\") pod \"apiserver-76f77b778f-46vzw\" (UID: \"b0ea3010-e975-4d1e-9fa4-385489660daa\") " pod="openshift-apiserver/apiserver-76f77b778f-46vzw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.573947 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.573985 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8ffe49e7-9382-4711-985c-8d6f76c42abc-installation-pull-secrets\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.574005 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.574023 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/82a8aff4-187a-464a-b375-5c20e576164d-auth-proxy-config\") pod \"machine-config-operator-74547568cd-gjtg8\" (UID: \"82a8aff4-187a-464a-b375-5c20e576164d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gjtg8" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.574040 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ffxh\" (UniqueName: \"kubernetes.io/projected/b160cc9e-253c-4aba-951a-21fd20ff52f4-kube-api-access-2ffxh\") pod \"downloads-7954f5f757-tc862\" (UID: \"b160cc9e-253c-4aba-951a-21fd20ff52f4\") " pod="openshift-console/downloads-7954f5f757-tc862" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.574160 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9t4mz\" (UniqueName: \"kubernetes.io/projected/57a3e335-0a9a-4fb5-b234-e0017b775ecd-kube-api-access-9t4mz\") pod \"kube-storage-version-migrator-operator-b67b599dd-dscrd\" (UID: \"57a3e335-0a9a-4fb5-b234-e0017b775ecd\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dscrd" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.574280 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0ea3010-e975-4d1e-9fa4-385489660daa-config\") pod \"apiserver-76f77b778f-46vzw\" (UID: \"b0ea3010-e975-4d1e-9fa4-385489660daa\") " pod="openshift-apiserver/apiserver-76f77b778f-46vzw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.574450 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/aa998110-e80e-4bc1-b1b2-f9806030ba26-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-jczvb\" (UID: \"aa998110-e80e-4bc1-b1b2-f9806030ba26\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jczvb" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.574728 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b0ea3010-e975-4d1e-9fa4-385489660daa-etcd-serving-ca\") pod \"apiserver-76f77b778f-46vzw\" (UID: \"b0ea3010-e975-4d1e-9fa4-385489660daa\") " pod="openshift-apiserver/apiserver-76f77b778f-46vzw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.574854 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-audit-policies\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.574928 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.575139 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8ffe49e7-9382-4711-985c-8d6f76c42abc-registry-certificates\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.575308 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b0ea3010-e975-4d1e-9fa4-385489660daa-trusted-ca-bundle\") pod \"apiserver-76f77b778f-46vzw\" (UID: \"b0ea3010-e975-4d1e-9fa4-385489660daa\") " pod="openshift-apiserver/apiserver-76f77b778f-46vzw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.575388 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8ffe49e7-9382-4711-985c-8d6f76c42abc-trusted-ca\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.575431 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b0ea3010-e975-4d1e-9fa4-385489660daa-audit-dir\") pod \"apiserver-76f77b778f-46vzw\" (UID: \"b0ea3010-e975-4d1e-9fa4-385489660daa\") " pod="openshift-apiserver/apiserver-76f77b778f-46vzw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.575466 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/652585ff-b553-4326-add1-ba20d0dd0a11-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-fzd5j\" (UID: \"652585ff-b553-4326-add1-ba20d0dd0a11\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fzd5j" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.591849 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.611327 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.632046 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.651563 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.672219 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.676151 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:27 crc kubenswrapper[4787]: E1001 09:38:27.676448 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:28.176421617 +0000 UTC m=+140.291565784 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.676495 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/e697b985-6356-4777-8ed2-06f9797a449c-certs\") pod \"machine-config-server-ns9tt\" (UID: \"e697b985-6356-4777-8ed2-06f9797a449c\") " pod="openshift-machine-config-operator/machine-config-server-ns9tt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.676533 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5d7d263b-bc04-4069-81a5-d5cd1151d805-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cf76m\" (UID: \"5d7d263b-bc04-4069-81a5-d5cd1151d805\") " pod="openshift-marketplace/marketplace-operator-79b997595-cf76m" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.676568 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8ffe49e7-9382-4711-985c-8d6f76c42abc-ca-trust-extracted\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.676614 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bvqz\" (UniqueName: \"kubernetes.io/projected/fa4ba99a-cfa6-4e3b-bd6f-ce18ea3fefe8-kube-api-access-6bvqz\") pod \"machine-config-controller-84d6567774-tzsx8\" (UID: \"fa4ba99a-cfa6-4e3b-bd6f-ce18ea3fefe8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tzsx8" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.676651 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzljp\" (UniqueName: \"kubernetes.io/projected/5d7d263b-bc04-4069-81a5-d5cd1151d805-kube-api-access-gzljp\") pod \"marketplace-operator-79b997595-cf76m\" (UID: \"5d7d263b-bc04-4069-81a5-d5cd1151d805\") " pod="openshift-marketplace/marketplace-operator-79b997595-cf76m" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.676683 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/151f124f-e071-4884-aeb4-217064807fa8-registration-dir\") pod \"csi-hostpathplugin-ncnkh\" (UID: \"151f124f-e071-4884-aeb4-217064807fa8\") " pod="hostpath-provisioner/csi-hostpathplugin-ncnkh" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.676707 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/151f124f-e071-4884-aeb4-217064807fa8-csi-data-dir\") pod \"csi-hostpathplugin-ncnkh\" (UID: \"151f124f-e071-4884-aeb4-217064807fa8\") " pod="hostpath-provisioner/csi-hostpathplugin-ncnkh" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.676844 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aa998110-e80e-4bc1-b1b2-f9806030ba26-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-jczvb\" (UID: \"aa998110-e80e-4bc1-b1b2-f9806030ba26\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jczvb" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.676878 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57a3e335-0a9a-4fb5-b234-e0017b775ecd-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-dscrd\" (UID: \"57a3e335-0a9a-4fb5-b234-e0017b775ecd\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dscrd" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.676904 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/13cafd2b-3d1e-4d8b-8751-a583af048423-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-blwqt\" (UID: \"13cafd2b-3d1e-4d8b-8751-a583af048423\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-blwqt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.677058 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4xvq\" (UniqueName: \"kubernetes.io/projected/e6b988c0-936d-4b49-80a2-015ae390f262-kube-api-access-t4xvq\") pod \"control-plane-machine-set-operator-78cbb6b69f-zssrz\" (UID: \"e6b988c0-936d-4b49-80a2-015ae390f262\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zssrz" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.677149 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8ffe49e7-9382-4711-985c-8d6f76c42abc-ca-trust-extracted\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.677236 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a2b762fa-551a-40c6-acd8-45e6649949d6-cert\") pod \"ingress-canary-sgxp6\" (UID: \"a2b762fa-551a-40c6-acd8-45e6649949d6\") " pod="openshift-ingress-canary/ingress-canary-sgxp6" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.677273 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e15a5f43-d273-42fb-b8d2-ba421fb1ab64-config\") pod \"service-ca-operator-777779d784-hqjh4\" (UID: \"e15a5f43-d273-42fb-b8d2-ba421fb1ab64\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hqjh4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.677302 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxgz5\" (UniqueName: \"kubernetes.io/projected/aca7d6cc-eca3-4ffb-84d9-254616d0c037-kube-api-access-gxgz5\") pod \"dns-operator-744455d44c-fsblz\" (UID: \"aca7d6cc-eca3-4ffb-84d9-254616d0c037\") " pod="openshift-dns-operator/dns-operator-744455d44c-fsblz" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.677402 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/54dbcab2-1757-4f7d-b59c-128e257b5660-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9x4nl\" (UID: \"54dbcab2-1757-4f7d-b59c-128e257b5660\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9x4nl" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.677538 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srs6n\" (UniqueName: \"kubernetes.io/projected/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-kube-api-access-srs6n\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.677621 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhq4f\" (UniqueName: \"kubernetes.io/projected/652585ff-b553-4326-add1-ba20d0dd0a11-kube-api-access-dhq4f\") pod \"openshift-controller-manager-operator-756b6f6bc6-fzd5j\" (UID: \"652585ff-b553-4326-add1-ba20d0dd0a11\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fzd5j" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.677647 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57a3e335-0a9a-4fb5-b234-e0017b775ecd-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-dscrd\" (UID: \"57a3e335-0a9a-4fb5-b234-e0017b775ecd\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dscrd" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.677705 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1d30c03-586b-460e-be72-c1770024c8f6-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-qvdw4\" (UID: \"e1d30c03-586b-460e-be72-c1770024c8f6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qvdw4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.677766 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/0ed372ed-3529-4e31-b124-d66f292089d0-machine-approver-tls\") pod \"machine-approver-56656f9798-pzdzn\" (UID: \"0ed372ed-3529-4e31-b124-d66f292089d0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pzdzn" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.677928 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b0ea3010-e975-4d1e-9fa4-385489660daa-etcd-client\") pod \"apiserver-76f77b778f-46vzw\" (UID: \"b0ea3010-e975-4d1e-9fa4-385489660daa\") " pod="openshift-apiserver/apiserver-76f77b778f-46vzw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.678655 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e15a5f43-d273-42fb-b8d2-ba421fb1ab64-serving-cert\") pod \"service-ca-operator-777779d784-hqjh4\" (UID: \"e15a5f43-d273-42fb-b8d2-ba421fb1ab64\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hqjh4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.678728 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5d7d263b-bc04-4069-81a5-d5cd1151d805-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cf76m\" (UID: \"5d7d263b-bc04-4069-81a5-d5cd1151d805\") " pod="openshift-marketplace/marketplace-operator-79b997595-cf76m" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.678807 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.679206 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aa998110-e80e-4bc1-b1b2-f9806030ba26-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-jczvb\" (UID: \"aa998110-e80e-4bc1-b1b2-f9806030ba26\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jczvb" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.679472 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7nkk\" (UniqueName: \"kubernetes.io/projected/151f124f-e071-4884-aeb4-217064807fa8-kube-api-access-v7nkk\") pod \"csi-hostpathplugin-ncnkh\" (UID: \"151f124f-e071-4884-aeb4-217064807fa8\") " pod="hostpath-provisioner/csi-hostpathplugin-ncnkh" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.679545 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.679607 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86g4f\" (UniqueName: \"kubernetes.io/projected/b0ea3010-e975-4d1e-9fa4-385489660daa-kube-api-access-86g4f\") pod \"apiserver-76f77b778f-46vzw\" (UID: \"b0ea3010-e975-4d1e-9fa4-385489660daa\") " pod="openshift-apiserver/apiserver-76f77b778f-46vzw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.679690 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlccc\" (UniqueName: \"kubernetes.io/projected/aa998110-e80e-4bc1-b1b2-f9806030ba26-kube-api-access-vlccc\") pod \"cluster-image-registry-operator-dc59b4c8b-jczvb\" (UID: \"aa998110-e80e-4bc1-b1b2-f9806030ba26\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jczvb" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.679764 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9338499-2f07-4087-b75f-bf85d8a2a794-config\") pod \"kube-controller-manager-operator-78b949d7b-86dvz\" (UID: \"e9338499-2f07-4087-b75f-bf85d8a2a794\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-86dvz" Oct 01 09:38:27 crc kubenswrapper[4787]: E1001 09:38:27.679954 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:28.179935366 +0000 UTC m=+140.295079753 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.680030 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjbhw\" (UniqueName: \"kubernetes.io/projected/82a8aff4-187a-464a-b375-5c20e576164d-kube-api-access-hjbhw\") pod \"machine-config-operator-74547568cd-gjtg8\" (UID: \"82a8aff4-187a-464a-b375-5c20e576164d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gjtg8" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.680798 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hr4xq\" (UniqueName: \"kubernetes.io/projected/75578ed4-3bb2-4ed7-bac1-d28780be4d67-kube-api-access-hr4xq\") pod \"multus-admission-controller-857f4d67dd-n6959\" (UID: \"75578ed4-3bb2-4ed7-bac1-d28780be4d67\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-n6959" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.680896 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/82a8aff4-187a-464a-b375-5c20e576164d-images\") pod \"machine-config-operator-74547568cd-gjtg8\" (UID: \"82a8aff4-187a-464a-b375-5c20e576164d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gjtg8" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.682171 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/e6b988c0-936d-4b49-80a2-015ae390f262-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zssrz\" (UID: \"e6b988c0-936d-4b49-80a2-015ae390f262\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zssrz" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.682212 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/151f124f-e071-4884-aeb4-217064807fa8-mountpoint-dir\") pod \"csi-hostpathplugin-ncnkh\" (UID: \"151f124f-e071-4884-aeb4-217064807fa8\") " pod="hostpath-provisioner/csi-hostpathplugin-ncnkh" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.681168 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b0ea3010-e975-4d1e-9fa4-385489660daa-etcd-client\") pod \"apiserver-76f77b778f-46vzw\" (UID: \"b0ea3010-e975-4d1e-9fa4-385489660daa\") " pod="openshift-apiserver/apiserver-76f77b778f-46vzw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.681288 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.682291 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/82a8aff4-187a-464a-b375-5c20e576164d-images\") pod \"machine-config-operator-74547568cd-gjtg8\" (UID: \"82a8aff4-187a-464a-b375-5c20e576164d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gjtg8" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.682387 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8ffe49e7-9382-4711-985c-8d6f76c42abc-registry-tls\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.682467 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fa4ba99a-cfa6-4e3b-bd6f-ce18ea3fefe8-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-tzsx8\" (UID: \"fa4ba99a-cfa6-4e3b-bd6f-ce18ea3fefe8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tzsx8" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.682571 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/556c69e8-1179-4de2-841d-3f096c3adb18-srv-cert\") pod \"olm-operator-6b444d44fb-whc4s\" (UID: \"556c69e8-1179-4de2-841d-3f096c3adb18\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-whc4s" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.682798 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0ed372ed-3529-4e31-b124-d66f292089d0-auth-proxy-config\") pod \"machine-approver-56656f9798-pzdzn\" (UID: \"0ed372ed-3529-4e31-b124-d66f292089d0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pzdzn" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.682880 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwmhz\" (UniqueName: \"kubernetes.io/projected/0ed372ed-3529-4e31-b124-d66f292089d0-kube-api-access-xwmhz\") pod \"machine-approver-56656f9798-pzdzn\" (UID: \"0ed372ed-3529-4e31-b124-d66f292089d0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pzdzn" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.682985 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c56c7aa-b794-474b-910b-11ecf8481009-config\") pod \"etcd-operator-b45778765-sgphp\" (UID: \"5c56c7aa-b794-474b-910b-11ecf8481009\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sgphp" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.683056 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfxzm\" (UniqueName: \"kubernetes.io/projected/13cafd2b-3d1e-4d8b-8751-a583af048423-kube-api-access-qfxzm\") pod \"package-server-manager-789f6589d5-blwqt\" (UID: \"13cafd2b-3d1e-4d8b-8751-a583af048423\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-blwqt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.683183 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4mj7\" (UniqueName: \"kubernetes.io/projected/59fc7ef6-b52d-41e5-8c82-9ce4ed5d5053-kube-api-access-d4mj7\") pod \"service-ca-9c57cc56f-tn48z\" (UID: \"59fc7ef6-b52d-41e5-8c82-9ce4ed5d5053\") " pod="openshift-service-ca/service-ca-9c57cc56f-tn48z" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.683258 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8ffe49e7-9382-4711-985c-8d6f76c42abc-installation-pull-secrets\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.683324 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.683394 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ffxh\" (UniqueName: \"kubernetes.io/projected/b160cc9e-253c-4aba-951a-21fd20ff52f4-kube-api-access-2ffxh\") pod \"downloads-7954f5f757-tc862\" (UID: \"b160cc9e-253c-4aba-951a-21fd20ff52f4\") " pod="openshift-console/downloads-7954f5f757-tc862" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.683527 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d438f633-0698-4c39-9c5a-71a57a6b3fbc-trusted-ca\") pod \"console-operator-58897d9998-hqx7l\" (UID: \"d438f633-0698-4c39-9c5a-71a57a6b3fbc\") " pod="openshift-console-operator/console-operator-58897d9998-hqx7l" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.683601 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/151f124f-e071-4884-aeb4-217064807fa8-plugins-dir\") pod \"csi-hostpathplugin-ncnkh\" (UID: \"151f124f-e071-4884-aeb4-217064807fa8\") " pod="hostpath-provisioner/csi-hostpathplugin-ncnkh" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.683729 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b0ea3010-e975-4d1e-9fa4-385489660daa-etcd-serving-ca\") pod \"apiserver-76f77b778f-46vzw\" (UID: \"b0ea3010-e975-4d1e-9fa4-385489660daa\") " pod="openshift-apiserver/apiserver-76f77b778f-46vzw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.683799 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-audit-policies\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.683963 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.684067 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0f1d5b79-c7bd-4c71-aa2c-7e407c4b7adb-trusted-ca\") pod \"ingress-operator-5b745b69d9-n47zx\" (UID: \"0f1d5b79-c7bd-4c71-aa2c-7e407c4b7adb\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n47zx" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.684187 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54dbcab2-1757-4f7d-b59c-128e257b5660-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9x4nl\" (UID: \"54dbcab2-1757-4f7d-b59c-128e257b5660\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9x4nl" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.684249 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8de4a645-2945-4261-b9be-ee3336e818df-secret-volume\") pod \"collect-profiles-29321850-kw99x\" (UID: \"8de4a645-2945-4261-b9be-ee3336e818df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-kw99x" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.684348 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8ffe49e7-9382-4711-985c-8d6f76c42abc-trusted-ca\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.684393 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5c56c7aa-b794-474b-910b-11ecf8481009-etcd-client\") pod \"etcd-operator-b45778765-sgphp\" (UID: \"5c56c7aa-b794-474b-910b-11ecf8481009\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sgphp" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.684416 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d438f633-0698-4c39-9c5a-71a57a6b3fbc-serving-cert\") pod \"console-operator-58897d9998-hqx7l\" (UID: \"d438f633-0698-4c39-9c5a-71a57a6b3fbc\") " pod="openshift-console-operator/console-operator-58897d9998-hqx7l" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.684472 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0f1d5b79-c7bd-4c71-aa2c-7e407c4b7adb-bound-sa-token\") pod \"ingress-operator-5b745b69d9-n47zx\" (UID: \"0f1d5b79-c7bd-4c71-aa2c-7e407c4b7adb\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n47zx" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.684515 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.684548 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/82a8aff4-187a-464a-b375-5c20e576164d-proxy-tls\") pod \"machine-config-operator-74547568cd-gjtg8\" (UID: \"82a8aff4-187a-464a-b375-5c20e576164d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gjtg8" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.684571 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgmrq\" (UniqueName: \"kubernetes.io/projected/8ffe49e7-9382-4711-985c-8d6f76c42abc-kube-api-access-dgmrq\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.684589 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2ec1df2f-eadd-4385-97b4-664101fb9d51-service-ca\") pod \"console-f9d7485db-gqvt9\" (UID: \"2ec1df2f-eadd-4385-97b4-664101fb9d51\") " pod="openshift-console/console-f9d7485db-gqvt9" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.684609 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/b0ea3010-e975-4d1e-9fa4-385489660daa-image-import-ca\") pod \"apiserver-76f77b778f-46vzw\" (UID: \"b0ea3010-e975-4d1e-9fa4-385489660daa\") " pod="openshift-apiserver/apiserver-76f77b778f-46vzw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.684634 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/aa998110-e80e-4bc1-b1b2-f9806030ba26-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-jczvb\" (UID: \"aa998110-e80e-4bc1-b1b2-f9806030ba26\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jczvb" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.684651 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b0ea3010-e975-4d1e-9fa4-385489660daa-node-pullsecrets\") pod \"apiserver-76f77b778f-46vzw\" (UID: \"b0ea3010-e975-4d1e-9fa4-385489660daa\") " pod="openshift-apiserver/apiserver-76f77b778f-46vzw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.684671 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/75578ed4-3bb2-4ed7-bac1-d28780be4d67-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-n6959\" (UID: \"75578ed4-3bb2-4ed7-bac1-d28780be4d67\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-n6959" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.684663 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fa4ba99a-cfa6-4e3b-bd6f-ce18ea3fefe8-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-tzsx8\" (UID: \"fa4ba99a-cfa6-4e3b-bd6f-ce18ea3fefe8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tzsx8" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.684688 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-px5h8\" (UniqueName: \"kubernetes.io/projected/9f2a2142-5430-4edb-8f20-b199d0425466-kube-api-access-px5h8\") pod \"dns-default-9lvh8\" (UID: \"9f2a2142-5430-4edb-8f20-b199d0425466\") " pod="openshift-dns/dns-default-9lvh8" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.684729 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9f2a2142-5430-4edb-8f20-b199d0425466-config-volume\") pod \"dns-default-9lvh8\" (UID: \"9f2a2142-5430-4edb-8f20-b199d0425466\") " pod="openshift-dns/dns-default-9lvh8" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.684745 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9lfw\" (UniqueName: \"kubernetes.io/projected/0f1d5b79-c7bd-4c71-aa2c-7e407c4b7adb-kube-api-access-h9lfw\") pod \"ingress-operator-5b745b69d9-n47zx\" (UID: \"0f1d5b79-c7bd-4c71-aa2c-7e407c4b7adb\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n47zx" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.684763 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2ec1df2f-eadd-4385-97b4-664101fb9d51-oauth-serving-cert\") pod \"console-f9d7485db-gqvt9\" (UID: \"2ec1df2f-eadd-4385-97b4-664101fb9d51\") " pod="openshift-console/console-f9d7485db-gqvt9" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.684788 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ed372ed-3529-4e31-b124-d66f292089d0-config\") pod \"machine-approver-56656f9798-pzdzn\" (UID: \"0ed372ed-3529-4e31-b124-d66f292089d0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pzdzn" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.684805 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/57a3e335-0a9a-4fb5-b234-e0017b775ecd-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-dscrd\" (UID: \"57a3e335-0a9a-4fb5-b234-e0017b775ecd\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dscrd" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.684822 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/ee393336-510d-40f3-a382-06d7f839750f-tmpfs\") pod \"packageserver-d55dfcdfc-r8xzr\" (UID: \"ee393336-510d-40f3-a382-06d7f839750f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r8xzr" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.684844 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.684859 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ee393336-510d-40f3-a382-06d7f839750f-apiservice-cert\") pod \"packageserver-d55dfcdfc-r8xzr\" (UID: \"ee393336-510d-40f3-a382-06d7f839750f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r8xzr" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.684880 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/59fc7ef6-b52d-41e5-8c82-9ce4ed5d5053-signing-key\") pod \"service-ca-9c57cc56f-tn48z\" (UID: \"59fc7ef6-b52d-41e5-8c82-9ce4ed5d5053\") " pod="openshift-service-ca/service-ca-9c57cc56f-tn48z" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.684905 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hflzd\" (UniqueName: \"kubernetes.io/projected/d438f633-0698-4c39-9c5a-71a57a6b3fbc-kube-api-access-hflzd\") pod \"console-operator-58897d9998-hqx7l\" (UID: \"d438f633-0698-4c39-9c5a-71a57a6b3fbc\") " pod="openshift-console-operator/console-operator-58897d9998-hqx7l" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.684931 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b0ea3010-e975-4d1e-9fa4-385489660daa-encryption-config\") pod \"apiserver-76f77b778f-46vzw\" (UID: \"b0ea3010-e975-4d1e-9fa4-385489660daa\") " pod="openshift-apiserver/apiserver-76f77b778f-46vzw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.684947 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d438f633-0698-4c39-9c5a-71a57a6b3fbc-config\") pod \"console-operator-58897d9998-hqx7l\" (UID: \"d438f633-0698-4c39-9c5a-71a57a6b3fbc\") " pod="openshift-console-operator/console-operator-58897d9998-hqx7l" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.684965 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/59fc7ef6-b52d-41e5-8c82-9ce4ed5d5053-signing-cabundle\") pod \"service-ca-9c57cc56f-tn48z\" (UID: \"59fc7ef6-b52d-41e5-8c82-9ce4ed5d5053\") " pod="openshift-service-ca/service-ca-9c57cc56f-tn48z" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.684988 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/aca7d6cc-eca3-4ffb-84d9-254616d0c037-metrics-tls\") pod \"dns-operator-744455d44c-fsblz\" (UID: \"aca7d6cc-eca3-4ffb-84d9-254616d0c037\") " pod="openshift-dns-operator/dns-operator-744455d44c-fsblz" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.685007 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49rkp\" (UniqueName: \"kubernetes.io/projected/5c56c7aa-b794-474b-910b-11ecf8481009-kube-api-access-49rkp\") pod \"etcd-operator-b45778765-sgphp\" (UID: \"5c56c7aa-b794-474b-910b-11ecf8481009\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sgphp" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.685029 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/98c8956e-c6c0-4e79-b7a5-64054256133a-available-featuregates\") pod \"openshift-config-operator-7777fb866f-lld4x\" (UID: \"98c8956e-c6c0-4e79-b7a5-64054256133a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-lld4x" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.685055 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/523e6f72-37ee-4c32-b3e0-98405775373f-config\") pod \"kube-apiserver-operator-766d6c64bb-29mhg\" (UID: \"523e6f72-37ee-4c32-b3e0-98405775373f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-29mhg" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.685071 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/11d20ebc-a959-43fa-84eb-bdc70a3f38f4-profile-collector-cert\") pod \"catalog-operator-68c6474976-rv6xx\" (UID: \"11d20ebc-a959-43fa-84eb-bdc70a3f38f4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rv6xx" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.685122 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2ec1df2f-eadd-4385-97b4-664101fb9d51-console-serving-cert\") pod \"console-f9d7485db-gqvt9\" (UID: \"2ec1df2f-eadd-4385-97b4-664101fb9d51\") " pod="openshift-console/console-f9d7485db-gqvt9" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.685143 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/652585ff-b553-4326-add1-ba20d0dd0a11-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-fzd5j\" (UID: \"652585ff-b553-4326-add1-ba20d0dd0a11\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fzd5j" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.685166 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4rsw\" (UniqueName: \"kubernetes.io/projected/e1d30c03-586b-460e-be72-c1770024c8f6-kube-api-access-s4rsw\") pod \"openshift-apiserver-operator-796bbdcf4f-qvdw4\" (UID: \"e1d30c03-586b-460e-be72-c1770024c8f6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qvdw4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.685184 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9smx9\" (UniqueName: \"kubernetes.io/projected/e15a5f43-d273-42fb-b8d2-ba421fb1ab64-kube-api-access-9smx9\") pod \"service-ca-operator-777779d784-hqjh4\" (UID: \"e15a5f43-d273-42fb-b8d2-ba421fb1ab64\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hqjh4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.685199 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2ec1df2f-eadd-4385-97b4-664101fb9d51-trusted-ca-bundle\") pod \"console-f9d7485db-gqvt9\" (UID: \"2ec1df2f-eadd-4385-97b4-664101fb9d51\") " pod="openshift-console/console-f9d7485db-gqvt9" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.685222 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwlbb\" (UniqueName: \"kubernetes.io/projected/98c8956e-c6c0-4e79-b7a5-64054256133a-kube-api-access-hwlbb\") pod \"openshift-config-operator-7777fb866f-lld4x\" (UID: \"98c8956e-c6c0-4e79-b7a5-64054256133a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-lld4x" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.685245 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.685261 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/523e6f72-37ee-4c32-b3e0-98405775373f-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-29mhg\" (UID: \"523e6f72-37ee-4c32-b3e0-98405775373f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-29mhg" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.685277 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t97mr\" (UniqueName: \"kubernetes.io/projected/556c69e8-1179-4de2-841d-3f096c3adb18-kube-api-access-t97mr\") pod \"olm-operator-6b444d44fb-whc4s\" (UID: \"556c69e8-1179-4de2-841d-3f096c3adb18\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-whc4s" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.685292 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1d30c03-586b-460e-be72-c1770024c8f6-config\") pod \"openshift-apiserver-operator-796bbdcf4f-qvdw4\" (UID: \"e1d30c03-586b-460e-be72-c1770024c8f6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qvdw4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.685322 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hvkf\" (UniqueName: \"kubernetes.io/projected/ee393336-510d-40f3-a382-06d7f839750f-kube-api-access-8hvkf\") pod \"packageserver-d55dfcdfc-r8xzr\" (UID: \"ee393336-510d-40f3-a382-06d7f839750f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r8xzr" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.685339 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/11d20ebc-a959-43fa-84eb-bdc70a3f38f4-srv-cert\") pod \"catalog-operator-68c6474976-rv6xx\" (UID: \"11d20ebc-a959-43fa-84eb-bdc70a3f38f4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rv6xx" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.685366 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.685390 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fa4ba99a-cfa6-4e3b-bd6f-ce18ea3fefe8-proxy-tls\") pod \"machine-config-controller-84d6567774-tzsx8\" (UID: \"fa4ba99a-cfa6-4e3b-bd6f-ce18ea3fefe8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tzsx8" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.685408 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/b0ea3010-e975-4d1e-9fa4-385489660daa-audit\") pod \"apiserver-76f77b778f-46vzw\" (UID: \"b0ea3010-e975-4d1e-9fa4-385489660daa\") " pod="openshift-apiserver/apiserver-76f77b778f-46vzw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.685430 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpsvj\" (UniqueName: \"kubernetes.io/projected/11d20ebc-a959-43fa-84eb-bdc70a3f38f4-kube-api-access-dpsvj\") pod \"catalog-operator-68c6474976-rv6xx\" (UID: \"11d20ebc-a959-43fa-84eb-bdc70a3f38f4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rv6xx" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.685456 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8de4a645-2945-4261-b9be-ee3336e818df-config-volume\") pod \"collect-profiles-29321850-kw99x\" (UID: \"8de4a645-2945-4261-b9be-ee3336e818df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-kw99x" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.685473 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/e697b985-6356-4777-8ed2-06f9797a449c-node-bootstrap-token\") pod \"machine-config-server-ns9tt\" (UID: \"e697b985-6356-4777-8ed2-06f9797a449c\") " pod="openshift-machine-config-operator/machine-config-server-ns9tt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.685490 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9338499-2f07-4087-b75f-bf85d8a2a794-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-86dvz\" (UID: \"e9338499-2f07-4087-b75f-bf85d8a2a794\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-86dvz" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.685513 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8ffe49e7-9382-4711-985c-8d6f76c42abc-bound-sa-token\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.685533 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0f1d5b79-c7bd-4c71-aa2c-7e407c4b7adb-metrics-tls\") pod \"ingress-operator-5b745b69d9-n47zx\" (UID: \"0f1d5b79-c7bd-4c71-aa2c-7e407c4b7adb\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n47zx" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.685551 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7f2hj\" (UniqueName: \"kubernetes.io/projected/2ec1df2f-eadd-4385-97b4-664101fb9d51-kube-api-access-7f2hj\") pod \"console-f9d7485db-gqvt9\" (UID: \"2ec1df2f-eadd-4385-97b4-664101fb9d51\") " pod="openshift-console/console-f9d7485db-gqvt9" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.684745 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b0ea3010-e975-4d1e-9fa4-385489660daa-etcd-serving-ca\") pod \"apiserver-76f77b778f-46vzw\" (UID: \"b0ea3010-e975-4d1e-9fa4-385489660daa\") " pod="openshift-apiserver/apiserver-76f77b778f-46vzw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.685576 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/556c69e8-1179-4de2-841d-3f096c3adb18-profile-collector-cert\") pod \"olm-operator-6b444d44fb-whc4s\" (UID: \"556c69e8-1179-4de2-841d-3f096c3adb18\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-whc4s" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.685626 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b0ea3010-e975-4d1e-9fa4-385489660daa-node-pullsecrets\") pod \"apiserver-76f77b778f-46vzw\" (UID: \"b0ea3010-e975-4d1e-9fa4-385489660daa\") " pod="openshift-apiserver/apiserver-76f77b778f-46vzw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.685655 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b0ea3010-e975-4d1e-9fa4-385489660daa-serving-cert\") pod \"apiserver-76f77b778f-46vzw\" (UID: \"b0ea3010-e975-4d1e-9fa4-385489660daa\") " pod="openshift-apiserver/apiserver-76f77b778f-46vzw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.685767 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-audit-policies\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.685796 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9f2a2142-5430-4edb-8f20-b199d0425466-metrics-tls\") pod \"dns-default-9lvh8\" (UID: \"9f2a2142-5430-4edb-8f20-b199d0425466\") " pod="openshift-dns/dns-default-9lvh8" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.685845 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.685826 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8ffe49e7-9382-4711-985c-8d6f76c42abc-registry-tls\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.686472 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.687900 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8ffe49e7-9382-4711-985c-8d6f76c42abc-trusted-ca\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.688067 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.688286 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c56c7aa-b794-474b-910b-11ecf8481009-serving-cert\") pod \"etcd-operator-b45778765-sgphp\" (UID: \"5c56c7aa-b794-474b-910b-11ecf8481009\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sgphp" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.688368 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dl48\" (UniqueName: \"kubernetes.io/projected/a2b762fa-551a-40c6-acd8-45e6649949d6-kube-api-access-6dl48\") pod \"ingress-canary-sgxp6\" (UID: \"a2b762fa-551a-40c6-acd8-45e6649949d6\") " pod="openshift-ingress-canary/ingress-canary-sgxp6" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.688414 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.688436 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2ec1df2f-eadd-4385-97b4-664101fb9d51-console-config\") pod \"console-f9d7485db-gqvt9\" (UID: \"2ec1df2f-eadd-4385-97b4-664101fb9d51\") " pod="openshift-console/console-f9d7485db-gqvt9" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.688543 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/82a8aff4-187a-464a-b375-5c20e576164d-auth-proxy-config\") pod \"machine-config-operator-74547568cd-gjtg8\" (UID: \"82a8aff4-187a-464a-b375-5c20e576164d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gjtg8" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.688580 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/652585ff-b553-4326-add1-ba20d0dd0a11-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-fzd5j\" (UID: \"652585ff-b553-4326-add1-ba20d0dd0a11\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fzd5j" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.688607 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/523e6f72-37ee-4c32-b3e0-98405775373f-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-29mhg\" (UID: \"523e6f72-37ee-4c32-b3e0-98405775373f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-29mhg" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.688672 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9t4mz\" (UniqueName: \"kubernetes.io/projected/57a3e335-0a9a-4fb5-b234-e0017b775ecd-kube-api-access-9t4mz\") pod \"kube-storage-version-migrator-operator-b67b599dd-dscrd\" (UID: \"57a3e335-0a9a-4fb5-b234-e0017b775ecd\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dscrd" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.688818 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0ea3010-e975-4d1e-9fa4-385489660daa-config\") pod \"apiserver-76f77b778f-46vzw\" (UID: \"b0ea3010-e975-4d1e-9fa4-385489660daa\") " pod="openshift-apiserver/apiserver-76f77b778f-46vzw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.688893 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/5c56c7aa-b794-474b-910b-11ecf8481009-etcd-ca\") pod \"etcd-operator-b45778765-sgphp\" (UID: \"5c56c7aa-b794-474b-910b-11ecf8481009\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sgphp" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.688949 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e9338499-2f07-4087-b75f-bf85d8a2a794-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-86dvz\" (UID: \"e9338499-2f07-4087-b75f-bf85d8a2a794\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-86dvz" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.689007 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2ec1df2f-eadd-4385-97b4-664101fb9d51-console-oauth-config\") pod \"console-f9d7485db-gqvt9\" (UID: \"2ec1df2f-eadd-4385-97b4-664101fb9d51\") " pod="openshift-console/console-f9d7485db-gqvt9" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.689201 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/aa998110-e80e-4bc1-b1b2-f9806030ba26-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-jczvb\" (UID: \"aa998110-e80e-4bc1-b1b2-f9806030ba26\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jczvb" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.689242 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.689281 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ee393336-510d-40f3-a382-06d7f839750f-webhook-cert\") pod \"packageserver-d55dfcdfc-r8xzr\" (UID: \"ee393336-510d-40f3-a382-06d7f839750f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r8xzr" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.689341 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/b0ea3010-e975-4d1e-9fa4-385489660daa-audit\") pod \"apiserver-76f77b778f-46vzw\" (UID: \"b0ea3010-e975-4d1e-9fa4-385489660daa\") " pod="openshift-apiserver/apiserver-76f77b778f-46vzw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.689361 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8ffe49e7-9382-4711-985c-8d6f76c42abc-installation-pull-secrets\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.689342 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8ffe49e7-9382-4711-985c-8d6f76c42abc-registry-certificates\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.689423 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtwkc\" (UniqueName: \"kubernetes.io/projected/e697b985-6356-4777-8ed2-06f9797a449c-kube-api-access-jtwkc\") pod \"machine-config-server-ns9tt\" (UID: \"e697b985-6356-4777-8ed2-06f9797a449c\") " pod="openshift-machine-config-operator/machine-config-server-ns9tt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.689470 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b0ea3010-e975-4d1e-9fa4-385489660daa-trusted-ca-bundle\") pod \"apiserver-76f77b778f-46vzw\" (UID: \"b0ea3010-e975-4d1e-9fa4-385489660daa\") " pod="openshift-apiserver/apiserver-76f77b778f-46vzw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.689503 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98c8956e-c6c0-4e79-b7a5-64054256133a-serving-cert\") pod \"openshift-config-operator-7777fb866f-lld4x\" (UID: \"98c8956e-c6c0-4e79-b7a5-64054256133a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-lld4x" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.689697 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0ea3010-e975-4d1e-9fa4-385489660daa-config\") pod \"apiserver-76f77b778f-46vzw\" (UID: \"b0ea3010-e975-4d1e-9fa4-385489660daa\") " pod="openshift-apiserver/apiserver-76f77b778f-46vzw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.689816 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/aa998110-e80e-4bc1-b1b2-f9806030ba26-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-jczvb\" (UID: \"aa998110-e80e-4bc1-b1b2-f9806030ba26\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jczvb" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.689935 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b0ea3010-e975-4d1e-9fa4-385489660daa-audit-dir\") pod \"apiserver-76f77b778f-46vzw\" (UID: \"b0ea3010-e975-4d1e-9fa4-385489660daa\") " pod="openshift-apiserver/apiserver-76f77b778f-46vzw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.690006 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/652585ff-b553-4326-add1-ba20d0dd0a11-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-fzd5j\" (UID: \"652585ff-b553-4326-add1-ba20d0dd0a11\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fzd5j" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.690234 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54dbcab2-1757-4f7d-b59c-128e257b5660-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9x4nl\" (UID: \"54dbcab2-1757-4f7d-b59c-128e257b5660\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9x4nl" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.690322 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.690364 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b0ea3010-e975-4d1e-9fa4-385489660daa-audit-dir\") pod \"apiserver-76f77b778f-46vzw\" (UID: \"b0ea3010-e975-4d1e-9fa4-385489660daa\") " pod="openshift-apiserver/apiserver-76f77b778f-46vzw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.690398 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.690329 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/82a8aff4-187a-464a-b375-5c20e576164d-auth-proxy-config\") pod \"machine-config-operator-74547568cd-gjtg8\" (UID: \"82a8aff4-187a-464a-b375-5c20e576164d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gjtg8" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.690645 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/151f124f-e071-4884-aeb4-217064807fa8-socket-dir\") pod \"csi-hostpathplugin-ncnkh\" (UID: \"151f124f-e071-4884-aeb4-217064807fa8\") " pod="hostpath-provisioner/csi-hostpathplugin-ncnkh" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.690741 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.690816 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dm6j\" (UniqueName: \"kubernetes.io/projected/8de4a645-2945-4261-b9be-ee3336e818df-kube-api-access-6dm6j\") pod \"collect-profiles-29321850-kw99x\" (UID: \"8de4a645-2945-4261-b9be-ee3336e818df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-kw99x" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.690652 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b0ea3010-e975-4d1e-9fa4-385489660daa-trusted-ca-bundle\") pod \"apiserver-76f77b778f-46vzw\" (UID: \"b0ea3010-e975-4d1e-9fa4-385489660daa\") " pod="openshift-apiserver/apiserver-76f77b778f-46vzw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.690975 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/5c56c7aa-b794-474b-910b-11ecf8481009-etcd-service-ca\") pod \"etcd-operator-b45778765-sgphp\" (UID: \"5c56c7aa-b794-474b-910b-11ecf8481009\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sgphp" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.691063 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-audit-dir\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.691393 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-audit-dir\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.691489 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b0ea3010-e975-4d1e-9fa4-385489660daa-serving-cert\") pod \"apiserver-76f77b778f-46vzw\" (UID: \"b0ea3010-e975-4d1e-9fa4-385489660daa\") " pod="openshift-apiserver/apiserver-76f77b778f-46vzw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.691771 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b0ea3010-e975-4d1e-9fa4-385489660daa-encryption-config\") pod \"apiserver-76f77b778f-46vzw\" (UID: \"b0ea3010-e975-4d1e-9fa4-385489660daa\") " pod="openshift-apiserver/apiserver-76f77b778f-46vzw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.691785 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.691862 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/57a3e335-0a9a-4fb5-b234-e0017b775ecd-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-dscrd\" (UID: \"57a3e335-0a9a-4fb5-b234-e0017b775ecd\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dscrd" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.692164 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8ffe49e7-9382-4711-985c-8d6f76c42abc-registry-certificates\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.692756 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fa4ba99a-cfa6-4e3b-bd6f-ce18ea3fefe8-proxy-tls\") pod \"machine-config-controller-84d6567774-tzsx8\" (UID: \"fa4ba99a-cfa6-4e3b-bd6f-ce18ea3fefe8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tzsx8" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.692768 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.693063 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/b0ea3010-e975-4d1e-9fa4-385489660daa-image-import-ca\") pod \"apiserver-76f77b778f-46vzw\" (UID: \"b0ea3010-e975-4d1e-9fa4-385489660daa\") " pod="openshift-apiserver/apiserver-76f77b778f-46vzw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.693576 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.693952 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.694550 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.694924 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.695000 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.695422 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/652585ff-b553-4326-add1-ba20d0dd0a11-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-fzd5j\" (UID: \"652585ff-b553-4326-add1-ba20d0dd0a11\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fzd5j" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.697912 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/82a8aff4-187a-464a-b375-5c20e576164d-proxy-tls\") pod \"machine-config-operator-74547568cd-gjtg8\" (UID: \"82a8aff4-187a-464a-b375-5c20e576164d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gjtg8" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.712008 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.731432 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.751375 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.770794 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.791818 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:27 crc kubenswrapper[4787]: E1001 09:38:27.792681 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:28.292027186 +0000 UTC m=+140.407171343 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.792736 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/523e6f72-37ee-4c32-b3e0-98405775373f-config\") pod \"kube-apiserver-operator-766d6c64bb-29mhg\" (UID: \"523e6f72-37ee-4c32-b3e0-98405775373f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-29mhg" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.792763 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/11d20ebc-a959-43fa-84eb-bdc70a3f38f4-profile-collector-cert\") pod \"catalog-operator-68c6474976-rv6xx\" (UID: \"11d20ebc-a959-43fa-84eb-bdc70a3f38f4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rv6xx" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.792783 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2ec1df2f-eadd-4385-97b4-664101fb9d51-console-serving-cert\") pod \"console-f9d7485db-gqvt9\" (UID: \"2ec1df2f-eadd-4385-97b4-664101fb9d51\") " pod="openshift-console/console-f9d7485db-gqvt9" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.793055 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwlbb\" (UniqueName: \"kubernetes.io/projected/98c8956e-c6c0-4e79-b7a5-64054256133a-kube-api-access-hwlbb\") pod \"openshift-config-operator-7777fb866f-lld4x\" (UID: \"98c8956e-c6c0-4e79-b7a5-64054256133a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-lld4x" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.793119 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4rsw\" (UniqueName: \"kubernetes.io/projected/e1d30c03-586b-460e-be72-c1770024c8f6-kube-api-access-s4rsw\") pod \"openshift-apiserver-operator-796bbdcf4f-qvdw4\" (UID: \"e1d30c03-586b-460e-be72-c1770024c8f6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qvdw4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.793140 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9smx9\" (UniqueName: \"kubernetes.io/projected/e15a5f43-d273-42fb-b8d2-ba421fb1ab64-kube-api-access-9smx9\") pod \"service-ca-operator-777779d784-hqjh4\" (UID: \"e15a5f43-d273-42fb-b8d2-ba421fb1ab64\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hqjh4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.793196 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2ec1df2f-eadd-4385-97b4-664101fb9d51-trusted-ca-bundle\") pod \"console-f9d7485db-gqvt9\" (UID: \"2ec1df2f-eadd-4385-97b4-664101fb9d51\") " pod="openshift-console/console-f9d7485db-gqvt9" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.793220 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t97mr\" (UniqueName: \"kubernetes.io/projected/556c69e8-1179-4de2-841d-3f096c3adb18-kube-api-access-t97mr\") pod \"olm-operator-6b444d44fb-whc4s\" (UID: \"556c69e8-1179-4de2-841d-3f096c3adb18\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-whc4s" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.793473 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1d30c03-586b-460e-be72-c1770024c8f6-config\") pod \"openshift-apiserver-operator-796bbdcf4f-qvdw4\" (UID: \"e1d30c03-586b-460e-be72-c1770024c8f6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qvdw4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.793495 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/523e6f72-37ee-4c32-b3e0-98405775373f-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-29mhg\" (UID: \"523e6f72-37ee-4c32-b3e0-98405775373f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-29mhg" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.793516 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hvkf\" (UniqueName: \"kubernetes.io/projected/ee393336-510d-40f3-a382-06d7f839750f-kube-api-access-8hvkf\") pod \"packageserver-d55dfcdfc-r8xzr\" (UID: \"ee393336-510d-40f3-a382-06d7f839750f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r8xzr" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.793554 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/11d20ebc-a959-43fa-84eb-bdc70a3f38f4-srv-cert\") pod \"catalog-operator-68c6474976-rv6xx\" (UID: \"11d20ebc-a959-43fa-84eb-bdc70a3f38f4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rv6xx" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.793622 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9338499-2f07-4087-b75f-bf85d8a2a794-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-86dvz\" (UID: \"e9338499-2f07-4087-b75f-bf85d8a2a794\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-86dvz" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.793641 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpsvj\" (UniqueName: \"kubernetes.io/projected/11d20ebc-a959-43fa-84eb-bdc70a3f38f4-kube-api-access-dpsvj\") pod \"catalog-operator-68c6474976-rv6xx\" (UID: \"11d20ebc-a959-43fa-84eb-bdc70a3f38f4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rv6xx" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.793678 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8de4a645-2945-4261-b9be-ee3336e818df-config-volume\") pod \"collect-profiles-29321850-kw99x\" (UID: \"8de4a645-2945-4261-b9be-ee3336e818df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-kw99x" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.793713 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/e697b985-6356-4777-8ed2-06f9797a449c-node-bootstrap-token\") pod \"machine-config-server-ns9tt\" (UID: \"e697b985-6356-4777-8ed2-06f9797a449c\") " pod="openshift-machine-config-operator/machine-config-server-ns9tt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.793739 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0f1d5b79-c7bd-4c71-aa2c-7e407c4b7adb-metrics-tls\") pod \"ingress-operator-5b745b69d9-n47zx\" (UID: \"0f1d5b79-c7bd-4c71-aa2c-7e407c4b7adb\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n47zx" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.793754 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7f2hj\" (UniqueName: \"kubernetes.io/projected/2ec1df2f-eadd-4385-97b4-664101fb9d51-kube-api-access-7f2hj\") pod \"console-f9d7485db-gqvt9\" (UID: \"2ec1df2f-eadd-4385-97b4-664101fb9d51\") " pod="openshift-console/console-f9d7485db-gqvt9" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.793856 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/556c69e8-1179-4de2-841d-3f096c3adb18-profile-collector-cert\") pod \"olm-operator-6b444d44fb-whc4s\" (UID: \"556c69e8-1179-4de2-841d-3f096c3adb18\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-whc4s" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.793877 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9f2a2142-5430-4edb-8f20-b199d0425466-metrics-tls\") pod \"dns-default-9lvh8\" (UID: \"9f2a2142-5430-4edb-8f20-b199d0425466\") " pod="openshift-dns/dns-default-9lvh8" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.793917 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/523e6f72-37ee-4c32-b3e0-98405775373f-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-29mhg\" (UID: \"523e6f72-37ee-4c32-b3e0-98405775373f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-29mhg" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.793935 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c56c7aa-b794-474b-910b-11ecf8481009-serving-cert\") pod \"etcd-operator-b45778765-sgphp\" (UID: \"5c56c7aa-b794-474b-910b-11ecf8481009\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sgphp" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.794887 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1d30c03-586b-460e-be72-c1770024c8f6-config\") pod \"openshift-apiserver-operator-796bbdcf4f-qvdw4\" (UID: \"e1d30c03-586b-460e-be72-c1770024c8f6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qvdw4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.795116 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dl48\" (UniqueName: \"kubernetes.io/projected/a2b762fa-551a-40c6-acd8-45e6649949d6-kube-api-access-6dl48\") pod \"ingress-canary-sgxp6\" (UID: \"a2b762fa-551a-40c6-acd8-45e6649949d6\") " pod="openshift-ingress-canary/ingress-canary-sgxp6" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.795328 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2ec1df2f-eadd-4385-97b4-664101fb9d51-console-config\") pod \"console-f9d7485db-gqvt9\" (UID: \"2ec1df2f-eadd-4385-97b4-664101fb9d51\") " pod="openshift-console/console-f9d7485db-gqvt9" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.795399 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/5c56c7aa-b794-474b-910b-11ecf8481009-etcd-ca\") pod \"etcd-operator-b45778765-sgphp\" (UID: \"5c56c7aa-b794-474b-910b-11ecf8481009\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sgphp" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.795467 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e9338499-2f07-4087-b75f-bf85d8a2a794-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-86dvz\" (UID: \"e9338499-2f07-4087-b75f-bf85d8a2a794\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-86dvz" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.795509 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2ec1df2f-eadd-4385-97b4-664101fb9d51-console-oauth-config\") pod \"console-f9d7485db-gqvt9\" (UID: \"2ec1df2f-eadd-4385-97b4-664101fb9d51\") " pod="openshift-console/console-f9d7485db-gqvt9" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.795551 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ee393336-510d-40f3-a382-06d7f839750f-webhook-cert\") pod \"packageserver-d55dfcdfc-r8xzr\" (UID: \"ee393336-510d-40f3-a382-06d7f839750f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r8xzr" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.795594 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtwkc\" (UniqueName: \"kubernetes.io/projected/e697b985-6356-4777-8ed2-06f9797a449c-kube-api-access-jtwkc\") pod \"machine-config-server-ns9tt\" (UID: \"e697b985-6356-4777-8ed2-06f9797a449c\") " pod="openshift-machine-config-operator/machine-config-server-ns9tt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.795645 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98c8956e-c6c0-4e79-b7a5-64054256133a-serving-cert\") pod \"openshift-config-operator-7777fb866f-lld4x\" (UID: \"98c8956e-c6c0-4e79-b7a5-64054256133a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-lld4x" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.795699 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54dbcab2-1757-4f7d-b59c-128e257b5660-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9x4nl\" (UID: \"54dbcab2-1757-4f7d-b59c-128e257b5660\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9x4nl" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.795769 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/151f124f-e071-4884-aeb4-217064807fa8-socket-dir\") pod \"csi-hostpathplugin-ncnkh\" (UID: \"151f124f-e071-4884-aeb4-217064807fa8\") " pod="hostpath-provisioner/csi-hostpathplugin-ncnkh" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.795809 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dm6j\" (UniqueName: \"kubernetes.io/projected/8de4a645-2945-4261-b9be-ee3336e818df-kube-api-access-6dm6j\") pod \"collect-profiles-29321850-kw99x\" (UID: \"8de4a645-2945-4261-b9be-ee3336e818df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-kw99x" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.795859 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/5c56c7aa-b794-474b-910b-11ecf8481009-etcd-service-ca\") pod \"etcd-operator-b45778765-sgphp\" (UID: \"5c56c7aa-b794-474b-910b-11ecf8481009\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sgphp" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.795902 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/e697b985-6356-4777-8ed2-06f9797a449c-certs\") pod \"machine-config-server-ns9tt\" (UID: \"e697b985-6356-4777-8ed2-06f9797a449c\") " pod="openshift-machine-config-operator/machine-config-server-ns9tt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.795948 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5d7d263b-bc04-4069-81a5-d5cd1151d805-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cf76m\" (UID: \"5d7d263b-bc04-4069-81a5-d5cd1151d805\") " pod="openshift-marketplace/marketplace-operator-79b997595-cf76m" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.795991 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/151f124f-e071-4884-aeb4-217064807fa8-registration-dir\") pod \"csi-hostpathplugin-ncnkh\" (UID: \"151f124f-e071-4884-aeb4-217064807fa8\") " pod="hostpath-provisioner/csi-hostpathplugin-ncnkh" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.796028 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/151f124f-e071-4884-aeb4-217064807fa8-csi-data-dir\") pod \"csi-hostpathplugin-ncnkh\" (UID: \"151f124f-e071-4884-aeb4-217064807fa8\") " pod="hostpath-provisioner/csi-hostpathplugin-ncnkh" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.796123 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/151f124f-e071-4884-aeb4-217064807fa8-socket-dir\") pod \"csi-hostpathplugin-ncnkh\" (UID: \"151f124f-e071-4884-aeb4-217064807fa8\") " pod="hostpath-provisioner/csi-hostpathplugin-ncnkh" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.796127 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/151f124f-e071-4884-aeb4-217064807fa8-registration-dir\") pod \"csi-hostpathplugin-ncnkh\" (UID: \"151f124f-e071-4884-aeb4-217064807fa8\") " pod="hostpath-provisioner/csi-hostpathplugin-ncnkh" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.796193 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/151f124f-e071-4884-aeb4-217064807fa8-csi-data-dir\") pod \"csi-hostpathplugin-ncnkh\" (UID: \"151f124f-e071-4884-aeb4-217064807fa8\") " pod="hostpath-provisioner/csi-hostpathplugin-ncnkh" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.796196 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzljp\" (UniqueName: \"kubernetes.io/projected/5d7d263b-bc04-4069-81a5-d5cd1151d805-kube-api-access-gzljp\") pod \"marketplace-operator-79b997595-cf76m\" (UID: \"5d7d263b-bc04-4069-81a5-d5cd1151d805\") " pod="openshift-marketplace/marketplace-operator-79b997595-cf76m" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.796289 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a2b762fa-551a-40c6-acd8-45e6649949d6-cert\") pod \"ingress-canary-sgxp6\" (UID: \"a2b762fa-551a-40c6-acd8-45e6649949d6\") " pod="openshift-ingress-canary/ingress-canary-sgxp6" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.796318 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/13cafd2b-3d1e-4d8b-8751-a583af048423-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-blwqt\" (UID: \"13cafd2b-3d1e-4d8b-8751-a583af048423\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-blwqt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.796346 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4xvq\" (UniqueName: \"kubernetes.io/projected/e6b988c0-936d-4b49-80a2-015ae390f262-kube-api-access-t4xvq\") pod \"control-plane-machine-set-operator-78cbb6b69f-zssrz\" (UID: \"e6b988c0-936d-4b49-80a2-015ae390f262\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zssrz" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.796373 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e15a5f43-d273-42fb-b8d2-ba421fb1ab64-config\") pod \"service-ca-operator-777779d784-hqjh4\" (UID: \"e15a5f43-d273-42fb-b8d2-ba421fb1ab64\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hqjh4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.796403 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxgz5\" (UniqueName: \"kubernetes.io/projected/aca7d6cc-eca3-4ffb-84d9-254616d0c037-kube-api-access-gxgz5\") pod \"dns-operator-744455d44c-fsblz\" (UID: \"aca7d6cc-eca3-4ffb-84d9-254616d0c037\") " pod="openshift-dns-operator/dns-operator-744455d44c-fsblz" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.796404 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54dbcab2-1757-4f7d-b59c-128e257b5660-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9x4nl\" (UID: \"54dbcab2-1757-4f7d-b59c-128e257b5660\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9x4nl" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.796426 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/54dbcab2-1757-4f7d-b59c-128e257b5660-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9x4nl\" (UID: \"54dbcab2-1757-4f7d-b59c-128e257b5660\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9x4nl" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.796480 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1d30c03-586b-460e-be72-c1770024c8f6-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-qvdw4\" (UID: \"e1d30c03-586b-460e-be72-c1770024c8f6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qvdw4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.796509 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/0ed372ed-3529-4e31-b124-d66f292089d0-machine-approver-tls\") pod \"machine-approver-56656f9798-pzdzn\" (UID: \"0ed372ed-3529-4e31-b124-d66f292089d0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pzdzn" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.796554 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5d7d263b-bc04-4069-81a5-d5cd1151d805-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cf76m\" (UID: \"5d7d263b-bc04-4069-81a5-d5cd1151d805\") " pod="openshift-marketplace/marketplace-operator-79b997595-cf76m" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.796586 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e15a5f43-d273-42fb-b8d2-ba421fb1ab64-serving-cert\") pod \"service-ca-operator-777779d784-hqjh4\" (UID: \"e15a5f43-d273-42fb-b8d2-ba421fb1ab64\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hqjh4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.796636 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.796679 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7nkk\" (UniqueName: \"kubernetes.io/projected/151f124f-e071-4884-aeb4-217064807fa8-kube-api-access-v7nkk\") pod \"csi-hostpathplugin-ncnkh\" (UID: \"151f124f-e071-4884-aeb4-217064807fa8\") " pod="hostpath-provisioner/csi-hostpathplugin-ncnkh" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.796712 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9338499-2f07-4087-b75f-bf85d8a2a794-config\") pod \"kube-controller-manager-operator-78b949d7b-86dvz\" (UID: \"e9338499-2f07-4087-b75f-bf85d8a2a794\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-86dvz" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.796753 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hr4xq\" (UniqueName: \"kubernetes.io/projected/75578ed4-3bb2-4ed7-bac1-d28780be4d67-kube-api-access-hr4xq\") pod \"multus-admission-controller-857f4d67dd-n6959\" (UID: \"75578ed4-3bb2-4ed7-bac1-d28780be4d67\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-n6959" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.796786 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/e6b988c0-936d-4b49-80a2-015ae390f262-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zssrz\" (UID: \"e6b988c0-936d-4b49-80a2-015ae390f262\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zssrz" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.796812 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/151f124f-e071-4884-aeb4-217064807fa8-mountpoint-dir\") pod \"csi-hostpathplugin-ncnkh\" (UID: \"151f124f-e071-4884-aeb4-217064807fa8\") " pod="hostpath-provisioner/csi-hostpathplugin-ncnkh" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.796839 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/556c69e8-1179-4de2-841d-3f096c3adb18-srv-cert\") pod \"olm-operator-6b444d44fb-whc4s\" (UID: \"556c69e8-1179-4de2-841d-3f096c3adb18\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-whc4s" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.796874 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0ed372ed-3529-4e31-b124-d66f292089d0-auth-proxy-config\") pod \"machine-approver-56656f9798-pzdzn\" (UID: \"0ed372ed-3529-4e31-b124-d66f292089d0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pzdzn" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.796890 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/151f124f-e071-4884-aeb4-217064807fa8-mountpoint-dir\") pod \"csi-hostpathplugin-ncnkh\" (UID: \"151f124f-e071-4884-aeb4-217064807fa8\") " pod="hostpath-provisioner/csi-hostpathplugin-ncnkh" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.796900 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwmhz\" (UniqueName: \"kubernetes.io/projected/0ed372ed-3529-4e31-b124-d66f292089d0-kube-api-access-xwmhz\") pod \"machine-approver-56656f9798-pzdzn\" (UID: \"0ed372ed-3529-4e31-b124-d66f292089d0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pzdzn" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.796929 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfxzm\" (UniqueName: \"kubernetes.io/projected/13cafd2b-3d1e-4d8b-8751-a583af048423-kube-api-access-qfxzm\") pod \"package-server-manager-789f6589d5-blwqt\" (UID: \"13cafd2b-3d1e-4d8b-8751-a583af048423\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-blwqt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.796956 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c56c7aa-b794-474b-910b-11ecf8481009-config\") pod \"etcd-operator-b45778765-sgphp\" (UID: \"5c56c7aa-b794-474b-910b-11ecf8481009\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sgphp" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.796986 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4mj7\" (UniqueName: \"kubernetes.io/projected/59fc7ef6-b52d-41e5-8c82-9ce4ed5d5053-kube-api-access-d4mj7\") pod \"service-ca-9c57cc56f-tn48z\" (UID: \"59fc7ef6-b52d-41e5-8c82-9ce4ed5d5053\") " pod="openshift-service-ca/service-ca-9c57cc56f-tn48z" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.797008 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d438f633-0698-4c39-9c5a-71a57a6b3fbc-trusted-ca\") pod \"console-operator-58897d9998-hqx7l\" (UID: \"d438f633-0698-4c39-9c5a-71a57a6b3fbc\") " pod="openshift-console-operator/console-operator-58897d9998-hqx7l" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.797031 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/151f124f-e071-4884-aeb4-217064807fa8-plugins-dir\") pod \"csi-hostpathplugin-ncnkh\" (UID: \"151f124f-e071-4884-aeb4-217064807fa8\") " pod="hostpath-provisioner/csi-hostpathplugin-ncnkh" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.797058 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 01 09:38:27 crc kubenswrapper[4787]: E1001 09:38:27.797216 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:28.297188468 +0000 UTC m=+140.412332665 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.797111 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54dbcab2-1757-4f7d-b59c-128e257b5660-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9x4nl\" (UID: \"54dbcab2-1757-4f7d-b59c-128e257b5660\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9x4nl" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.797399 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0f1d5b79-c7bd-4c71-aa2c-7e407c4b7adb-trusted-ca\") pod \"ingress-operator-5b745b69d9-n47zx\" (UID: \"0f1d5b79-c7bd-4c71-aa2c-7e407c4b7adb\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n47zx" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.797437 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8de4a645-2945-4261-b9be-ee3336e818df-secret-volume\") pod \"collect-profiles-29321850-kw99x\" (UID: \"8de4a645-2945-4261-b9be-ee3336e818df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-kw99x" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.797461 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0f1d5b79-c7bd-4c71-aa2c-7e407c4b7adb-bound-sa-token\") pod \"ingress-operator-5b745b69d9-n47zx\" (UID: \"0f1d5b79-c7bd-4c71-aa2c-7e407c4b7adb\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n47zx" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.797502 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5c56c7aa-b794-474b-910b-11ecf8481009-etcd-client\") pod \"etcd-operator-b45778765-sgphp\" (UID: \"5c56c7aa-b794-474b-910b-11ecf8481009\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sgphp" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.797505 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/151f124f-e071-4884-aeb4-217064807fa8-plugins-dir\") pod \"csi-hostpathplugin-ncnkh\" (UID: \"151f124f-e071-4884-aeb4-217064807fa8\") " pod="hostpath-provisioner/csi-hostpathplugin-ncnkh" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.797527 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d438f633-0698-4c39-9c5a-71a57a6b3fbc-serving-cert\") pod \"console-operator-58897d9998-hqx7l\" (UID: \"d438f633-0698-4c39-9c5a-71a57a6b3fbc\") " pod="openshift-console-operator/console-operator-58897d9998-hqx7l" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.797537 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0ed372ed-3529-4e31-b124-d66f292089d0-auth-proxy-config\") pod \"machine-approver-56656f9798-pzdzn\" (UID: \"0ed372ed-3529-4e31-b124-d66f292089d0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pzdzn" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.797581 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2ec1df2f-eadd-4385-97b4-664101fb9d51-service-ca\") pod \"console-f9d7485db-gqvt9\" (UID: \"2ec1df2f-eadd-4385-97b4-664101fb9d51\") " pod="openshift-console/console-f9d7485db-gqvt9" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.797612 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/75578ed4-3bb2-4ed7-bac1-d28780be4d67-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-n6959\" (UID: \"75578ed4-3bb2-4ed7-bac1-d28780be4d67\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-n6959" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.797638 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-px5h8\" (UniqueName: \"kubernetes.io/projected/9f2a2142-5430-4edb-8f20-b199d0425466-kube-api-access-px5h8\") pod \"dns-default-9lvh8\" (UID: \"9f2a2142-5430-4edb-8f20-b199d0425466\") " pod="openshift-dns/dns-default-9lvh8" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.797688 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9f2a2142-5430-4edb-8f20-b199d0425466-config-volume\") pod \"dns-default-9lvh8\" (UID: \"9f2a2142-5430-4edb-8f20-b199d0425466\") " pod="openshift-dns/dns-default-9lvh8" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.797711 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9lfw\" (UniqueName: \"kubernetes.io/projected/0f1d5b79-c7bd-4c71-aa2c-7e407c4b7adb-kube-api-access-h9lfw\") pod \"ingress-operator-5b745b69d9-n47zx\" (UID: \"0f1d5b79-c7bd-4c71-aa2c-7e407c4b7adb\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n47zx" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.797738 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2ec1df2f-eadd-4385-97b4-664101fb9d51-oauth-serving-cert\") pod \"console-f9d7485db-gqvt9\" (UID: \"2ec1df2f-eadd-4385-97b4-664101fb9d51\") " pod="openshift-console/console-f9d7485db-gqvt9" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.797763 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/ee393336-510d-40f3-a382-06d7f839750f-tmpfs\") pod \"packageserver-d55dfcdfc-r8xzr\" (UID: \"ee393336-510d-40f3-a382-06d7f839750f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r8xzr" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.797787 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ed372ed-3529-4e31-b124-d66f292089d0-config\") pod \"machine-approver-56656f9798-pzdzn\" (UID: \"0ed372ed-3529-4e31-b124-d66f292089d0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pzdzn" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.797810 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/59fc7ef6-b52d-41e5-8c82-9ce4ed5d5053-signing-key\") pod \"service-ca-9c57cc56f-tn48z\" (UID: \"59fc7ef6-b52d-41e5-8c82-9ce4ed5d5053\") " pod="openshift-service-ca/service-ca-9c57cc56f-tn48z" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.797837 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ee393336-510d-40f3-a382-06d7f839750f-apiservice-cert\") pod \"packageserver-d55dfcdfc-r8xzr\" (UID: \"ee393336-510d-40f3-a382-06d7f839750f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r8xzr" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.797875 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hflzd\" (UniqueName: \"kubernetes.io/projected/d438f633-0698-4c39-9c5a-71a57a6b3fbc-kube-api-access-hflzd\") pod \"console-operator-58897d9998-hqx7l\" (UID: \"d438f633-0698-4c39-9c5a-71a57a6b3fbc\") " pod="openshift-console-operator/console-operator-58897d9998-hqx7l" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.797906 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d438f633-0698-4c39-9c5a-71a57a6b3fbc-config\") pod \"console-operator-58897d9998-hqx7l\" (UID: \"d438f633-0698-4c39-9c5a-71a57a6b3fbc\") " pod="openshift-console-operator/console-operator-58897d9998-hqx7l" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.797929 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/59fc7ef6-b52d-41e5-8c82-9ce4ed5d5053-signing-cabundle\") pod \"service-ca-9c57cc56f-tn48z\" (UID: \"59fc7ef6-b52d-41e5-8c82-9ce4ed5d5053\") " pod="openshift-service-ca/service-ca-9c57cc56f-tn48z" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.797957 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/aca7d6cc-eca3-4ffb-84d9-254616d0c037-metrics-tls\") pod \"dns-operator-744455d44c-fsblz\" (UID: \"aca7d6cc-eca3-4ffb-84d9-254616d0c037\") " pod="openshift-dns-operator/dns-operator-744455d44c-fsblz" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.797982 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49rkp\" (UniqueName: \"kubernetes.io/projected/5c56c7aa-b794-474b-910b-11ecf8481009-kube-api-access-49rkp\") pod \"etcd-operator-b45778765-sgphp\" (UID: \"5c56c7aa-b794-474b-910b-11ecf8481009\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sgphp" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.798008 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/98c8956e-c6c0-4e79-b7a5-64054256133a-available-featuregates\") pod \"openshift-config-operator-7777fb866f-lld4x\" (UID: \"98c8956e-c6c0-4e79-b7a5-64054256133a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-lld4x" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.798063 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0f1d5b79-c7bd-4c71-aa2c-7e407c4b7adb-metrics-tls\") pod \"ingress-operator-5b745b69d9-n47zx\" (UID: \"0f1d5b79-c7bd-4c71-aa2c-7e407c4b7adb\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n47zx" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.798378 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98c8956e-c6c0-4e79-b7a5-64054256133a-serving-cert\") pod \"openshift-config-operator-7777fb866f-lld4x\" (UID: \"98c8956e-c6c0-4e79-b7a5-64054256133a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-lld4x" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.798567 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/98c8956e-c6c0-4e79-b7a5-64054256133a-available-featuregates\") pod \"openshift-config-operator-7777fb866f-lld4x\" (UID: \"98c8956e-c6c0-4e79-b7a5-64054256133a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-lld4x" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.798641 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0f1d5b79-c7bd-4c71-aa2c-7e407c4b7adb-trusted-ca\") pod \"ingress-operator-5b745b69d9-n47zx\" (UID: \"0f1d5b79-c7bd-4c71-aa2c-7e407c4b7adb\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n47zx" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.798640 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/ee393336-510d-40f3-a382-06d7f839750f-tmpfs\") pod \"packageserver-d55dfcdfc-r8xzr\" (UID: \"ee393336-510d-40f3-a382-06d7f839750f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r8xzr" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.799371 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ed372ed-3529-4e31-b124-d66f292089d0-config\") pod \"machine-approver-56656f9798-pzdzn\" (UID: \"0ed372ed-3529-4e31-b124-d66f292089d0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pzdzn" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.799464 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d438f633-0698-4c39-9c5a-71a57a6b3fbc-trusted-ca\") pod \"console-operator-58897d9998-hqx7l\" (UID: \"d438f633-0698-4c39-9c5a-71a57a6b3fbc\") " pod="openshift-console-operator/console-operator-58897d9998-hqx7l" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.799846 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d438f633-0698-4c39-9c5a-71a57a6b3fbc-config\") pod \"console-operator-58897d9998-hqx7l\" (UID: \"d438f633-0698-4c39-9c5a-71a57a6b3fbc\") " pod="openshift-console-operator/console-operator-58897d9998-hqx7l" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.800571 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1d30c03-586b-460e-be72-c1770024c8f6-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-qvdw4\" (UID: \"e1d30c03-586b-460e-be72-c1770024c8f6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qvdw4" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.801594 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d438f633-0698-4c39-9c5a-71a57a6b3fbc-serving-cert\") pod \"console-operator-58897d9998-hqx7l\" (UID: \"d438f633-0698-4c39-9c5a-71a57a6b3fbc\") " pod="openshift-console-operator/console-operator-58897d9998-hqx7l" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.802636 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54dbcab2-1757-4f7d-b59c-128e257b5660-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9x4nl\" (UID: \"54dbcab2-1757-4f7d-b59c-128e257b5660\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9x4nl" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.802784 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/aca7d6cc-eca3-4ffb-84d9-254616d0c037-metrics-tls\") pod \"dns-operator-744455d44c-fsblz\" (UID: \"aca7d6cc-eca3-4ffb-84d9-254616d0c037\") " pod="openshift-dns-operator/dns-operator-744455d44c-fsblz" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.803850 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/0ed372ed-3529-4e31-b124-d66f292089d0-machine-approver-tls\") pod \"machine-approver-56656f9798-pzdzn\" (UID: \"0ed372ed-3529-4e31-b124-d66f292089d0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pzdzn" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.810753 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.830955 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.851864 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.872401 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.877643 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/523e6f72-37ee-4c32-b3e0-98405775373f-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-29mhg\" (UID: \"523e6f72-37ee-4c32-b3e0-98405775373f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-29mhg" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.892527 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.894498 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/523e6f72-37ee-4c32-b3e0-98405775373f-config\") pod \"kube-apiserver-operator-766d6c64bb-29mhg\" (UID: \"523e6f72-37ee-4c32-b3e0-98405775373f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-29mhg" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.898951 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:27 crc kubenswrapper[4787]: E1001 09:38:27.899110 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:28.399094617 +0000 UTC m=+140.514238784 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.899714 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:27 crc kubenswrapper[4787]: E1001 09:38:27.900105 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:28.400094723 +0000 UTC m=+140.515238890 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.923872 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbcfn\" (UniqueName: \"kubernetes.io/projected/fad32d96-ad78-48b4-9874-bef9a0a30f8f-kube-api-access-xbcfn\") pod \"authentication-operator-69f744f599-cdc2f\" (UID: \"fad32d96-ad78-48b4-9874-bef9a0a30f8f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cdc2f" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.930848 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.950957 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.972608 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.979703 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/11d20ebc-a959-43fa-84eb-bdc70a3f38f4-srv-cert\") pod \"catalog-operator-68c6474976-rv6xx\" (UID: \"11d20ebc-a959-43fa-84eb-bdc70a3f38f4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rv6xx" Oct 01 09:38:27 crc kubenswrapper[4787]: I1001 09:38:27.991264 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.000437 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:28 crc kubenswrapper[4787]: E1001 09:38:28.000679 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:28.500636556 +0000 UTC m=+140.615780713 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.001469 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:28 crc kubenswrapper[4787]: E1001 09:38:28.001876 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:28.501861147 +0000 UTC m=+140.617005304 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.026577 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2pgd\" (UniqueName: \"kubernetes.io/projected/07892f22-556b-49a6-8f71-3d3b16b2cdef-kube-api-access-t2pgd\") pod \"machine-api-operator-5694c8668f-4g5xc\" (UID: \"07892f22-556b-49a6-8f71-3d3b16b2cdef\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4g5xc" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.044470 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89wcv\" (UniqueName: \"kubernetes.io/projected/5f699a8f-ffe8-4512-9e55-24c88e4154d2-kube-api-access-89wcv\") pod \"route-controller-manager-6576b87f9c-djtn2\" (UID: \"5f699a8f-ffe8-4512-9e55-24c88e4154d2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-djtn2" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.044788 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-cdc2f" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.065399 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-628nj\" (UniqueName: \"kubernetes.io/projected/e6f1b5a2-581a-4c1a-b952-5662b2fb636f-kube-api-access-628nj\") pod \"controller-manager-879f6c89f-x649j\" (UID: \"e6f1b5a2-581a-4c1a-b952-5662b2fb636f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-x649j" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.086957 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frkgv\" (UniqueName: \"kubernetes.io/projected/160b7ce2-3861-4c15-af76-2b0d41ec8301-kube-api-access-frkgv\") pod \"apiserver-7bbb656c7d-hlhp4\" (UID: \"160b7ce2-3861-4c15-af76-2b0d41ec8301\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhp4" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.092001 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.102407 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:28 crc kubenswrapper[4787]: E1001 09:38:28.102526 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:28.602493524 +0000 UTC m=+140.717637681 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.102712 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.102901 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/556c69e8-1179-4de2-841d-3f096c3adb18-srv-cert\") pod \"olm-operator-6b444d44fb-whc4s\" (UID: \"556c69e8-1179-4de2-841d-3f096c3adb18\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-whc4s" Oct 01 09:38:28 crc kubenswrapper[4787]: E1001 09:38:28.103129 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:28.6031191 +0000 UTC m=+140.718263367 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.112226 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.131125 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.152554 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.154300 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9338499-2f07-4087-b75f-bf85d8a2a794-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-86dvz\" (UID: \"e9338499-2f07-4087-b75f-bf85d8a2a794\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-86dvz" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.158262 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9338499-2f07-4087-b75f-bf85d8a2a794-config\") pod \"kube-controller-manager-operator-78b949d7b-86dvz\" (UID: \"e9338499-2f07-4087-b75f-bf85d8a2a794\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-86dvz" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.172319 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.192274 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.198634 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/11d20ebc-a959-43fa-84eb-bdc70a3f38f4-profile-collector-cert\") pod \"catalog-operator-68c6474976-rv6xx\" (UID: \"11d20ebc-a959-43fa-84eb-bdc70a3f38f4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rv6xx" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.198988 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/556c69e8-1179-4de2-841d-3f096c3adb18-profile-collector-cert\") pod \"olm-operator-6b444d44fb-whc4s\" (UID: \"556c69e8-1179-4de2-841d-3f096c3adb18\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-whc4s" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.204601 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:28 crc kubenswrapper[4787]: E1001 09:38:28.205276 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:28.705258965 +0000 UTC m=+140.820403122 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.205369 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:28 crc kubenswrapper[4787]: E1001 09:38:28.205864 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:28.70585478 +0000 UTC m=+140.820998937 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.207845 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8de4a645-2945-4261-b9be-ee3336e818df-secret-volume\") pod \"collect-profiles-29321850-kw99x\" (UID: \"8de4a645-2945-4261-b9be-ee3336e818df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-kw99x" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.212371 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.219860 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ee393336-510d-40f3-a382-06d7f839750f-webhook-cert\") pod \"packageserver-d55dfcdfc-r8xzr\" (UID: \"ee393336-510d-40f3-a382-06d7f839750f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r8xzr" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.221688 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ee393336-510d-40f3-a382-06d7f839750f-apiservice-cert\") pod \"packageserver-d55dfcdfc-r8xzr\" (UID: \"ee393336-510d-40f3-a382-06d7f839750f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r8xzr" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.231880 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.241054 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/75578ed4-3bb2-4ed7-bac1-d28780be4d67-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-n6959\" (UID: \"75578ed4-3bb2-4ed7-bac1-d28780be4d67\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-n6959" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.251039 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.276170 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-djtn2" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.276318 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.280778 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/e6b988c0-936d-4b49-80a2-015ae390f262-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zssrz\" (UID: \"e6b988c0-936d-4b49-80a2-015ae390f262\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zssrz" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.283704 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-cdc2f"] Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.289232 4787 request.go:700] Waited for 1.006965358s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-api/secrets?fieldSelector=metadata.name%3Dcontrol-plane-machine-set-operator-dockercfg-k9rxt&limit=500&resourceVersion=0 Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.290987 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.306780 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:28 crc kubenswrapper[4787]: E1001 09:38:28.307398 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:28.807377299 +0000 UTC m=+140.922521456 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.308004 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:28 crc kubenswrapper[4787]: E1001 09:38:28.308539 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:28.808530708 +0000 UTC m=+140.923674865 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.311881 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-4g5xc" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.312530 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.314961 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8de4a645-2945-4261-b9be-ee3336e818df-config-volume\") pod \"collect-profiles-29321850-kw99x\" (UID: \"8de4a645-2945-4261-b9be-ee3336e818df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-kw99x" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.322717 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhp4" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.332152 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.336970 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-x649j" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.351318 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.363447 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/59fc7ef6-b52d-41e5-8c82-9ce4ed5d5053-signing-key\") pod \"service-ca-9c57cc56f-tn48z\" (UID: \"59fc7ef6-b52d-41e5-8c82-9ce4ed5d5053\") " pod="openshift-service-ca/service-ca-9c57cc56f-tn48z" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.371281 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.392794 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.412458 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.412691 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 01 09:38:28 crc kubenswrapper[4787]: E1001 09:38:28.413494 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:28.913459724 +0000 UTC m=+141.028603881 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.421119 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/59fc7ef6-b52d-41e5-8c82-9ce4ed5d5053-signing-cabundle\") pod \"service-ca-9c57cc56f-tn48z\" (UID: \"59fc7ef6-b52d-41e5-8c82-9ce4ed5d5053\") " pod="openshift-service-ca/service-ca-9c57cc56f-tn48z" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.431994 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.451646 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.461261 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2ec1df2f-eadd-4385-97b4-664101fb9d51-console-oauth-config\") pod \"console-f9d7485db-gqvt9\" (UID: \"2ec1df2f-eadd-4385-97b4-664101fb9d51\") " pod="openshift-console/console-f9d7485db-gqvt9" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.471200 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.479878 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-djtn2"] Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.493029 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.497464 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2ec1df2f-eadd-4385-97b4-664101fb9d51-console-serving-cert\") pod \"console-f9d7485db-gqvt9\" (UID: \"2ec1df2f-eadd-4385-97b4-664101fb9d51\") " pod="openshift-console/console-f9d7485db-gqvt9" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.513677 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.514493 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:28 crc kubenswrapper[4787]: E1001 09:38:28.515183 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:29.015172999 +0000 UTC m=+141.130317156 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.522378 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2ec1df2f-eadd-4385-97b4-664101fb9d51-oauth-serving-cert\") pod \"console-f9d7485db-gqvt9\" (UID: \"2ec1df2f-eadd-4385-97b4-664101fb9d51\") " pod="openshift-console/console-f9d7485db-gqvt9" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.531729 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.538737 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2ec1df2f-eadd-4385-97b4-664101fb9d51-service-ca\") pod \"console-f9d7485db-gqvt9\" (UID: \"2ec1df2f-eadd-4385-97b4-664101fb9d51\") " pod="openshift-console/console-f9d7485db-gqvt9" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.548574 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhp4"] Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.551538 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.556187 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2ec1df2f-eadd-4385-97b4-664101fb9d51-console-config\") pod \"console-f9d7485db-gqvt9\" (UID: \"2ec1df2f-eadd-4385-97b4-664101fb9d51\") " pod="openshift-console/console-f9d7485db-gqvt9" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.556574 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-4g5xc"] Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.578717 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.582596 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-x649j"] Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.585559 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2ec1df2f-eadd-4385-97b4-664101fb9d51-trusted-ca-bundle\") pod \"console-f9d7485db-gqvt9\" (UID: \"2ec1df2f-eadd-4385-97b4-664101fb9d51\") " pod="openshift-console/console-f9d7485db-gqvt9" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.591296 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.601020 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/13cafd2b-3d1e-4d8b-8751-a583af048423-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-blwqt\" (UID: \"13cafd2b-3d1e-4d8b-8751-a583af048423\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-blwqt" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.611403 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.615338 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:28 crc kubenswrapper[4787]: E1001 09:38:28.615509 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:29.115491977 +0000 UTC m=+141.230636134 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.615705 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:28 crc kubenswrapper[4787]: E1001 09:38:28.616162 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:29.116145423 +0000 UTC m=+141.231289580 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.619220 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c56c7aa-b794-474b-910b-11ecf8481009-serving-cert\") pod \"etcd-operator-b45778765-sgphp\" (UID: \"5c56c7aa-b794-474b-910b-11ecf8481009\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sgphp" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.632061 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.652200 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.664421 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/5c56c7aa-b794-474b-910b-11ecf8481009-etcd-service-ca\") pod \"etcd-operator-b45778765-sgphp\" (UID: \"5c56c7aa-b794-474b-910b-11ecf8481009\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sgphp" Oct 01 09:38:28 crc kubenswrapper[4787]: W1001 09:38:28.667709 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod160b7ce2_3861_4c15_af76_2b0d41ec8301.slice/crio-3a6ae68a085a0845feff72c7c48bacab3579dd87b616788d131a5f9c4f40f423 WatchSource:0}: Error finding container 3a6ae68a085a0845feff72c7c48bacab3579dd87b616788d131a5f9c4f40f423: Status 404 returned error can't find the container with id 3a6ae68a085a0845feff72c7c48bacab3579dd87b616788d131a5f9c4f40f423 Oct 01 09:38:28 crc kubenswrapper[4787]: W1001 09:38:28.670263 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07892f22_556b_49a6_8f71_3d3b16b2cdef.slice/crio-96e2ab91455e63f54588a7386763b23f1a891618eb7eac6dd65871509bbefedc WatchSource:0}: Error finding container 96e2ab91455e63f54588a7386763b23f1a891618eb7eac6dd65871509bbefedc: Status 404 returned error can't find the container with id 96e2ab91455e63f54588a7386763b23f1a891618eb7eac6dd65871509bbefedc Oct 01 09:38:28 crc kubenswrapper[4787]: W1001 09:38:28.672033 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode6f1b5a2_581a_4c1a_b952_5662b2fb636f.slice/crio-476b72cc982d783e926c5708376f78357f203fd47d3076351abd671992867c3d WatchSource:0}: Error finding container 476b72cc982d783e926c5708376f78357f203fd47d3076351abd671992867c3d: Status 404 returned error can't find the container with id 476b72cc982d783e926c5708376f78357f203fd47d3076351abd671992867c3d Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.672331 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.682760 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5c56c7aa-b794-474b-910b-11ecf8481009-etcd-client\") pod \"etcd-operator-b45778765-sgphp\" (UID: \"5c56c7aa-b794-474b-910b-11ecf8481009\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sgphp" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.691237 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.712042 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.716828 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:28 crc kubenswrapper[4787]: E1001 09:38:28.716969 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:29.216944824 +0000 UTC m=+141.332088981 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.717110 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:28 crc kubenswrapper[4787]: E1001 09:38:28.717717 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:29.217678982 +0000 UTC m=+141.332823139 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.731664 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.736243 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/5c56c7aa-b794-474b-910b-11ecf8481009-etcd-ca\") pod \"etcd-operator-b45778765-sgphp\" (UID: \"5c56c7aa-b794-474b-910b-11ecf8481009\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sgphp" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.751582 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.757776 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c56c7aa-b794-474b-910b-11ecf8481009-config\") pod \"etcd-operator-b45778765-sgphp\" (UID: \"5c56c7aa-b794-474b-910b-11ecf8481009\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sgphp" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.791170 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 01 09:38:28 crc kubenswrapper[4787]: E1001 09:38:28.794403 4787 secret.go:188] Couldn't get secret openshift-dns/dns-default-metrics-tls: failed to sync secret cache: timed out waiting for the condition Oct 01 09:38:28 crc kubenswrapper[4787]: E1001 09:38:28.794431 4787 secret.go:188] Couldn't get secret openshift-machine-config-operator/node-bootstrapper-token: failed to sync secret cache: timed out waiting for the condition Oct 01 09:38:28 crc kubenswrapper[4787]: E1001 09:38:28.794472 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9f2a2142-5430-4edb-8f20-b199d0425466-metrics-tls podName:9f2a2142-5430-4edb-8f20-b199d0425466 nodeName:}" failed. No retries permitted until 2025-10-01 09:38:29.294452288 +0000 UTC m=+141.409596445 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/9f2a2142-5430-4edb-8f20-b199d0425466-metrics-tls") pod "dns-default-9lvh8" (UID: "9f2a2142-5430-4edb-8f20-b199d0425466") : failed to sync secret cache: timed out waiting for the condition Oct 01 09:38:28 crc kubenswrapper[4787]: E1001 09:38:28.794487 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e697b985-6356-4777-8ed2-06f9797a449c-node-bootstrap-token podName:e697b985-6356-4777-8ed2-06f9797a449c nodeName:}" failed. No retries permitted until 2025-10-01 09:38:29.294481768 +0000 UTC m=+141.409625925 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-bootstrap-token" (UniqueName: "kubernetes.io/secret/e697b985-6356-4777-8ed2-06f9797a449c-node-bootstrap-token") pod "machine-config-server-ns9tt" (UID: "e697b985-6356-4777-8ed2-06f9797a449c") : failed to sync secret cache: timed out waiting for the condition Oct 01 09:38:28 crc kubenswrapper[4787]: E1001 09:38:28.796689 4787 secret.go:188] Couldn't get secret openshift-marketplace/marketplace-operator-metrics: failed to sync secret cache: timed out waiting for the condition Oct 01 09:38:28 crc kubenswrapper[4787]: E1001 09:38:28.796722 4787 secret.go:188] Couldn't get secret openshift-machine-config-operator/machine-config-server-tls: failed to sync secret cache: timed out waiting for the condition Oct 01 09:38:28 crc kubenswrapper[4787]: E1001 09:38:28.796728 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5d7d263b-bc04-4069-81a5-d5cd1151d805-marketplace-operator-metrics podName:5d7d263b-bc04-4069-81a5-d5cd1151d805 nodeName:}" failed. No retries permitted until 2025-10-01 09:38:29.296720916 +0000 UTC m=+141.411865073 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-operator-metrics" (UniqueName: "kubernetes.io/secret/5d7d263b-bc04-4069-81a5-d5cd1151d805-marketplace-operator-metrics") pod "marketplace-operator-79b997595-cf76m" (UID: "5d7d263b-bc04-4069-81a5-d5cd1151d805") : failed to sync secret cache: timed out waiting for the condition Oct 01 09:38:28 crc kubenswrapper[4787]: E1001 09:38:28.796783 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e697b985-6356-4777-8ed2-06f9797a449c-certs podName:e697b985-6356-4777-8ed2-06f9797a449c nodeName:}" failed. No retries permitted until 2025-10-01 09:38:29.296772107 +0000 UTC m=+141.411916264 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "certs" (UniqueName: "kubernetes.io/secret/e697b985-6356-4777-8ed2-06f9797a449c-certs") pod "machine-config-server-ns9tt" (UID: "e697b985-6356-4777-8ed2-06f9797a449c") : failed to sync secret cache: timed out waiting for the condition Oct 01 09:38:28 crc kubenswrapper[4787]: E1001 09:38:28.796805 4787 configmap.go:193] Couldn't get configMap openshift-marketplace/marketplace-trusted-ca: failed to sync configmap cache: timed out waiting for the condition Oct 01 09:38:28 crc kubenswrapper[4787]: E1001 09:38:28.796805 4787 secret.go:188] Couldn't get secret openshift-service-ca-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 01 09:38:28 crc kubenswrapper[4787]: E1001 09:38:28.796836 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5d7d263b-bc04-4069-81a5-d5cd1151d805-marketplace-trusted-ca podName:5d7d263b-bc04-4069-81a5-d5cd1151d805 nodeName:}" failed. No retries permitted until 2025-10-01 09:38:29.296827369 +0000 UTC m=+141.411971526 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-trusted-ca" (UniqueName: "kubernetes.io/configmap/5d7d263b-bc04-4069-81a5-d5cd1151d805-marketplace-trusted-ca") pod "marketplace-operator-79b997595-cf76m" (UID: "5d7d263b-bc04-4069-81a5-d5cd1151d805") : failed to sync configmap cache: timed out waiting for the condition Oct 01 09:38:28 crc kubenswrapper[4787]: E1001 09:38:28.796933 4787 configmap.go:193] Couldn't get configMap openshift-service-ca-operator/service-ca-operator-config: failed to sync configmap cache: timed out waiting for the condition Oct 01 09:38:28 crc kubenswrapper[4787]: E1001 09:38:28.796830 4787 secret.go:188] Couldn't get secret openshift-ingress-canary/canary-serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 01 09:38:28 crc kubenswrapper[4787]: E1001 09:38:28.796972 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e15a5f43-d273-42fb-b8d2-ba421fb1ab64-serving-cert podName:e15a5f43-d273-42fb-b8d2-ba421fb1ab64 nodeName:}" failed. No retries permitted until 2025-10-01 09:38:29.296939382 +0000 UTC m=+141.412083569 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/e15a5f43-d273-42fb-b8d2-ba421fb1ab64-serving-cert") pod "service-ca-operator-777779d784-hqjh4" (UID: "e15a5f43-d273-42fb-b8d2-ba421fb1ab64") : failed to sync secret cache: timed out waiting for the condition Oct 01 09:38:28 crc kubenswrapper[4787]: E1001 09:38:28.797003 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e15a5f43-d273-42fb-b8d2-ba421fb1ab64-config podName:e15a5f43-d273-42fb-b8d2-ba421fb1ab64 nodeName:}" failed. No retries permitted until 2025-10-01 09:38:29.296990323 +0000 UTC m=+141.412134520 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/e15a5f43-d273-42fb-b8d2-ba421fb1ab64-config") pod "service-ca-operator-777779d784-hqjh4" (UID: "e15a5f43-d273-42fb-b8d2-ba421fb1ab64") : failed to sync configmap cache: timed out waiting for the condition Oct 01 09:38:28 crc kubenswrapper[4787]: E1001 09:38:28.797031 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a2b762fa-551a-40c6-acd8-45e6649949d6-cert podName:a2b762fa-551a-40c6-acd8-45e6649949d6 nodeName:}" failed. No retries permitted until 2025-10-01 09:38:29.297018204 +0000 UTC m=+141.412162401 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a2b762fa-551a-40c6-acd8-45e6649949d6-cert") pod "ingress-canary-sgxp6" (UID: "a2b762fa-551a-40c6-acd8-45e6649949d6") : failed to sync secret cache: timed out waiting for the condition Oct 01 09:38:28 crc kubenswrapper[4787]: E1001 09:38:28.798233 4787 configmap.go:193] Couldn't get configMap openshift-dns/dns-default: failed to sync configmap cache: timed out waiting for the condition Oct 01 09:38:28 crc kubenswrapper[4787]: E1001 09:38:28.798293 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/9f2a2142-5430-4edb-8f20-b199d0425466-config-volume podName:9f2a2142-5430-4edb-8f20-b199d0425466 nodeName:}" failed. No retries permitted until 2025-10-01 09:38:29.298279276 +0000 UTC m=+141.413423433 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/9f2a2142-5430-4edb-8f20-b199d0425466-config-volume") pod "dns-default-9lvh8" (UID: "9f2a2142-5430-4edb-8f20-b199d0425466") : failed to sync configmap cache: timed out waiting for the condition Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.811442 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.818917 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:28 crc kubenswrapper[4787]: E1001 09:38:28.819133 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:29.319113819 +0000 UTC m=+141.434257976 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.819934 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:28 crc kubenswrapper[4787]: E1001 09:38:28.820345 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:29.320333921 +0000 UTC m=+141.435478078 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.832476 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.851159 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.871711 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.892112 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.911869 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.939945 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.942017 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:28 crc kubenswrapper[4787]: E1001 09:38:28.942442 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:29.442417746 +0000 UTC m=+141.557561903 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.943189 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:28 crc kubenswrapper[4787]: E1001 09:38:28.944213 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:29.44416076 +0000 UTC m=+141.559304937 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.959953 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.971758 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 01 09:38:28 crc kubenswrapper[4787]: I1001 09:38:28.991998 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.012778 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.031602 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.044888 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:29 crc kubenswrapper[4787]: E1001 09:38:29.045523 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:29.545506405 +0000 UTC m=+141.660650562 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.068034 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzwr2\" (UniqueName: \"kubernetes.io/projected/52a18549-2c1c-4f28-bfa1-6fb14e0690e1-kube-api-access-hzwr2\") pod \"router-default-5444994796-7p4xb\" (UID: \"52a18549-2c1c-4f28-bfa1-6fb14e0690e1\") " pod="openshift-ingress/router-default-5444994796-7p4xb" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.085723 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmwmk\" (UniqueName: \"kubernetes.io/projected/1c07acd7-0397-494c-ab52-83bcdefdf2ac-kube-api-access-mmwmk\") pod \"cluster-samples-operator-665b6dd947-tvvw6\" (UID: \"1c07acd7-0397-494c-ab52-83bcdefdf2ac\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tvvw6" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.091037 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.110144 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-7p4xb" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.111987 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.131784 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.146934 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:29 crc kubenswrapper[4787]: E1001 09:38:29.147447 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:29.647425284 +0000 UTC m=+141.762569491 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:29 crc kubenswrapper[4787]: W1001 09:38:29.157266 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod52a18549_2c1c_4f28_bfa1_6fb14e0690e1.slice/crio-329ce7e0e158de8fd0bad12084fb85dcbe24edebcfcaf186684e72655ffe723e WatchSource:0}: Error finding container 329ce7e0e158de8fd0bad12084fb85dcbe24edebcfcaf186684e72655ffe723e: Status 404 returned error can't find the container with id 329ce7e0e158de8fd0bad12084fb85dcbe24edebcfcaf186684e72655ffe723e Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.165218 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k84xq\" (UniqueName: \"kubernetes.io/projected/15f1cb19-94ee-49b2-a837-b7bdaf484387-kube-api-access-k84xq\") pod \"migrator-59844c95c7-t2mbp\" (UID: \"15f1cb19-94ee-49b2-a837-b7bdaf484387\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t2mbp" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.172100 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.192107 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.212012 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.232156 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.238297 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-cdc2f" event={"ID":"fad32d96-ad78-48b4-9874-bef9a0a30f8f","Type":"ContainerStarted","Data":"5a78bd340fff8f884f534b4b2258ff7244e7ffd26cf241e52be09d567313e551"} Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.238354 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-cdc2f" event={"ID":"fad32d96-ad78-48b4-9874-bef9a0a30f8f","Type":"ContainerStarted","Data":"f559f3882e3dc33b4cb3acbe3448710716fc470a901a6a621674c089ea6d4ac5"} Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.240100 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-4g5xc" event={"ID":"07892f22-556b-49a6-8f71-3d3b16b2cdef","Type":"ContainerStarted","Data":"f696c52fa125d7b00ecc7a335666b92ce5ad6e0245a068818adb4ba497a692e9"} Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.240151 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-4g5xc" event={"ID":"07892f22-556b-49a6-8f71-3d3b16b2cdef","Type":"ContainerStarted","Data":"8e109d78b34995a0f7fd2dc7d3515d6873b7477a05c5b9326b9094aa6014f4b0"} Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.240166 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-4g5xc" event={"ID":"07892f22-556b-49a6-8f71-3d3b16b2cdef","Type":"ContainerStarted","Data":"96e2ab91455e63f54588a7386763b23f1a891618eb7eac6dd65871509bbefedc"} Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.241655 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-djtn2" event={"ID":"5f699a8f-ffe8-4512-9e55-24c88e4154d2","Type":"ContainerStarted","Data":"89335eacca626ba7a51ed255e2110f680e83ab9eef027d90d08b322e5ee08fd8"} Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.241684 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-djtn2" event={"ID":"5f699a8f-ffe8-4512-9e55-24c88e4154d2","Type":"ContainerStarted","Data":"63ef6ecd235c2be02d075ee4eb6afdff48eeb3978126c398915cca97cfb38c7a"} Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.241872 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-djtn2" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.243182 4787 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-djtn2 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.243240 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-djtn2" podUID="5f699a8f-ffe8-4512-9e55-24c88e4154d2" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.243398 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-x649j" event={"ID":"e6f1b5a2-581a-4c1a-b952-5662b2fb636f","Type":"ContainerStarted","Data":"6d26013c5f8421594cca1529739132a4ce30cbc1f627f8e35431996d17074ba8"} Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.243427 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-x649j" event={"ID":"e6f1b5a2-581a-4c1a-b952-5662b2fb636f","Type":"ContainerStarted","Data":"476b72cc982d783e926c5708376f78357f203fd47d3076351abd671992867c3d"} Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.243654 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-x649j" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.245027 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-7p4xb" event={"ID":"52a18549-2c1c-4f28-bfa1-6fb14e0690e1","Type":"ContainerStarted","Data":"329ce7e0e158de8fd0bad12084fb85dcbe24edebcfcaf186684e72655ffe723e"} Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.246275 4787 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-x649j container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.246308 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-x649j" podUID="e6f1b5a2-581a-4c1a-b952-5662b2fb636f" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.247615 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.247639 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhp4" event={"ID":"160b7ce2-3861-4c15-af76-2b0d41ec8301","Type":"ContainerDied","Data":"9913a944c50d6ecbd48f2c14cfe43e82ea9bcd028c3abb3374e93228820ad4a9"} Oct 01 09:38:29 crc kubenswrapper[4787]: E1001 09:38:29.247741 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:29.747716902 +0000 UTC m=+141.862861069 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.247614 4787 generic.go:334] "Generic (PLEG): container finished" podID="160b7ce2-3861-4c15-af76-2b0d41ec8301" containerID="9913a944c50d6ecbd48f2c14cfe43e82ea9bcd028c3abb3374e93228820ad4a9" exitCode=0 Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.247783 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhp4" event={"ID":"160b7ce2-3861-4c15-af76-2b0d41ec8301","Type":"ContainerStarted","Data":"3a6ae68a085a0845feff72c7c48bacab3579dd87b616788d131a5f9c4f40f423"} Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.248221 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:29 crc kubenswrapper[4787]: E1001 09:38:29.248552 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:29.748539363 +0000 UTC m=+141.863683520 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.252300 4787 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.256557 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tvvw6" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.272259 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.289343 4787 request.go:700] Waited for 1.898068781s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/hostpath-provisioner/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.292608 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.348378 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bvqz\" (UniqueName: \"kubernetes.io/projected/fa4ba99a-cfa6-4e3b-bd6f-ce18ea3fefe8-kube-api-access-6bvqz\") pod \"machine-config-controller-84d6567774-tzsx8\" (UID: \"fa4ba99a-cfa6-4e3b-bd6f-ce18ea3fefe8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tzsx8" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.348819 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:29 crc kubenswrapper[4787]: E1001 09:38:29.348994 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:29.848953033 +0000 UTC m=+141.964097200 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.349220 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9f2a2142-5430-4edb-8f20-b199d0425466-config-volume\") pod \"dns-default-9lvh8\" (UID: \"9f2a2142-5430-4edb-8f20-b199d0425466\") " pod="openshift-dns/dns-default-9lvh8" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.351916 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9f2a2142-5430-4edb-8f20-b199d0425466-config-volume\") pod \"dns-default-9lvh8\" (UID: \"9f2a2142-5430-4edb-8f20-b199d0425466\") " pod="openshift-dns/dns-default-9lvh8" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.352791 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/e697b985-6356-4777-8ed2-06f9797a449c-node-bootstrap-token\") pod \"machine-config-server-ns9tt\" (UID: \"e697b985-6356-4777-8ed2-06f9797a449c\") " pod="openshift-machine-config-operator/machine-config-server-ns9tt" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.352881 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9f2a2142-5430-4edb-8f20-b199d0425466-metrics-tls\") pod \"dns-default-9lvh8\" (UID: \"9f2a2142-5430-4edb-8f20-b199d0425466\") " pod="openshift-dns/dns-default-9lvh8" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.353048 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/e697b985-6356-4777-8ed2-06f9797a449c-certs\") pod \"machine-config-server-ns9tt\" (UID: \"e697b985-6356-4777-8ed2-06f9797a449c\") " pod="openshift-machine-config-operator/machine-config-server-ns9tt" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.353153 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5d7d263b-bc04-4069-81a5-d5cd1151d805-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cf76m\" (UID: \"5d7d263b-bc04-4069-81a5-d5cd1151d805\") " pod="openshift-marketplace/marketplace-operator-79b997595-cf76m" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.353231 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a2b762fa-551a-40c6-acd8-45e6649949d6-cert\") pod \"ingress-canary-sgxp6\" (UID: \"a2b762fa-551a-40c6-acd8-45e6649949d6\") " pod="openshift-ingress-canary/ingress-canary-sgxp6" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.353287 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e15a5f43-d273-42fb-b8d2-ba421fb1ab64-config\") pod \"service-ca-operator-777779d784-hqjh4\" (UID: \"e15a5f43-d273-42fb-b8d2-ba421fb1ab64\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hqjh4" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.353631 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e15a5f43-d273-42fb-b8d2-ba421fb1ab64-serving-cert\") pod \"service-ca-operator-777779d784-hqjh4\" (UID: \"e15a5f43-d273-42fb-b8d2-ba421fb1ab64\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hqjh4" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.353708 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5d7d263b-bc04-4069-81a5-d5cd1151d805-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cf76m\" (UID: \"5d7d263b-bc04-4069-81a5-d5cd1151d805\") " pod="openshift-marketplace/marketplace-operator-79b997595-cf76m" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.353755 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.354524 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e15a5f43-d273-42fb-b8d2-ba421fb1ab64-config\") pod \"service-ca-operator-777779d784-hqjh4\" (UID: \"e15a5f43-d273-42fb-b8d2-ba421fb1ab64\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hqjh4" Oct 01 09:38:29 crc kubenswrapper[4787]: E1001 09:38:29.374717 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:29.874670772 +0000 UTC m=+141.989814919 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.376378 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9f2a2142-5430-4edb-8f20-b199d0425466-metrics-tls\") pod \"dns-default-9lvh8\" (UID: \"9f2a2142-5430-4edb-8f20-b199d0425466\") " pod="openshift-dns/dns-default-9lvh8" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.376498 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srs6n\" (UniqueName: \"kubernetes.io/projected/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-kube-api-access-srs6n\") pod \"oauth-openshift-558db77b4-hwftw\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.378431 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e15a5f43-d273-42fb-b8d2-ba421fb1ab64-serving-cert\") pod \"service-ca-operator-777779d784-hqjh4\" (UID: \"e15a5f43-d273-42fb-b8d2-ba421fb1ab64\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hqjh4" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.380722 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5d7d263b-bc04-4069-81a5-d5cd1151d805-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cf76m\" (UID: \"5d7d263b-bc04-4069-81a5-d5cd1151d805\") " pod="openshift-marketplace/marketplace-operator-79b997595-cf76m" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.381755 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5d7d263b-bc04-4069-81a5-d5cd1151d805-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cf76m\" (UID: \"5d7d263b-bc04-4069-81a5-d5cd1151d805\") " pod="openshift-marketplace/marketplace-operator-79b997595-cf76m" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.383817 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhq4f\" (UniqueName: \"kubernetes.io/projected/652585ff-b553-4326-add1-ba20d0dd0a11-kube-api-access-dhq4f\") pod \"openshift-controller-manager-operator-756b6f6bc6-fzd5j\" (UID: \"652585ff-b553-4326-add1-ba20d0dd0a11\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fzd5j" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.384331 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/e697b985-6356-4777-8ed2-06f9797a449c-node-bootstrap-token\") pod \"machine-config-server-ns9tt\" (UID: \"e697b985-6356-4777-8ed2-06f9797a449c\") " pod="openshift-machine-config-operator/machine-config-server-ns9tt" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.385697 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/e697b985-6356-4777-8ed2-06f9797a449c-certs\") pod \"machine-config-server-ns9tt\" (UID: \"e697b985-6356-4777-8ed2-06f9797a449c\") " pod="openshift-machine-config-operator/machine-config-server-ns9tt" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.391999 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a2b762fa-551a-40c6-acd8-45e6649949d6-cert\") pod \"ingress-canary-sgxp6\" (UID: \"a2b762fa-551a-40c6-acd8-45e6649949d6\") " pod="openshift-ingress-canary/ingress-canary-sgxp6" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.397933 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86g4f\" (UniqueName: \"kubernetes.io/projected/b0ea3010-e975-4d1e-9fa4-385489660daa-kube-api-access-86g4f\") pod \"apiserver-76f77b778f-46vzw\" (UID: \"b0ea3010-e975-4d1e-9fa4-385489660daa\") " pod="openshift-apiserver/apiserver-76f77b778f-46vzw" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.410655 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlccc\" (UniqueName: \"kubernetes.io/projected/aa998110-e80e-4bc1-b1b2-f9806030ba26-kube-api-access-vlccc\") pod \"cluster-image-registry-operator-dc59b4c8b-jczvb\" (UID: \"aa998110-e80e-4bc1-b1b2-f9806030ba26\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jczvb" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.430218 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjbhw\" (UniqueName: \"kubernetes.io/projected/82a8aff4-187a-464a-b375-5c20e576164d-kube-api-access-hjbhw\") pod \"machine-config-operator-74547568cd-gjtg8\" (UID: \"82a8aff4-187a-464a-b375-5c20e576164d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gjtg8" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.439476 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t2mbp" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.447853 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ffxh\" (UniqueName: \"kubernetes.io/projected/b160cc9e-253c-4aba-951a-21fd20ff52f4-kube-api-access-2ffxh\") pod \"downloads-7954f5f757-tc862\" (UID: \"b160cc9e-253c-4aba-951a-21fd20ff52f4\") " pod="openshift-console/downloads-7954f5f757-tc862" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.449657 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-46vzw" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.460483 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fzd5j" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.462433 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:29 crc kubenswrapper[4787]: E1001 09:38:29.462595 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:29.962557811 +0000 UTC m=+142.077701968 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.463186 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:29 crc kubenswrapper[4787]: E1001 09:38:29.463592 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:29.963585088 +0000 UTC m=+142.078729245 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.466689 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tvvw6"] Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.481201 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgmrq\" (UniqueName: \"kubernetes.io/projected/8ffe49e7-9382-4711-985c-8d6f76c42abc-kube-api-access-dgmrq\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.493142 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8ffe49e7-9382-4711-985c-8d6f76c42abc-bound-sa-token\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.517835 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/aa998110-e80e-4bc1-b1b2-f9806030ba26-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-jczvb\" (UID: \"aa998110-e80e-4bc1-b1b2-f9806030ba26\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jczvb" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.531534 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9t4mz\" (UniqueName: \"kubernetes.io/projected/57a3e335-0a9a-4fb5-b234-e0017b775ecd-kube-api-access-9t4mz\") pod \"kube-storage-version-migrator-operator-b67b599dd-dscrd\" (UID: \"57a3e335-0a9a-4fb5-b234-e0017b775ecd\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dscrd" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.554605 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwlbb\" (UniqueName: \"kubernetes.io/projected/98c8956e-c6c0-4e79-b7a5-64054256133a-kube-api-access-hwlbb\") pod \"openshift-config-operator-7777fb866f-lld4x\" (UID: \"98c8956e-c6c0-4e79-b7a5-64054256133a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-lld4x" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.564408 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:29 crc kubenswrapper[4787]: E1001 09:38:29.565383 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:30.065346963 +0000 UTC m=+142.180491120 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.570769 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4rsw\" (UniqueName: \"kubernetes.io/projected/e1d30c03-586b-460e-be72-c1770024c8f6-kube-api-access-s4rsw\") pod \"openshift-apiserver-operator-796bbdcf4f-qvdw4\" (UID: \"e1d30c03-586b-460e-be72-c1770024c8f6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qvdw4" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.591006 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tzsx8" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.594792 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9smx9\" (UniqueName: \"kubernetes.io/projected/e15a5f43-d273-42fb-b8d2-ba421fb1ab64-kube-api-access-9smx9\") pod \"service-ca-operator-777779d784-hqjh4\" (UID: \"e15a5f43-d273-42fb-b8d2-ba421fb1ab64\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hqjh4" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.618797 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t97mr\" (UniqueName: \"kubernetes.io/projected/556c69e8-1179-4de2-841d-3f096c3adb18-kube-api-access-t97mr\") pod \"olm-operator-6b444d44fb-whc4s\" (UID: \"556c69e8-1179-4de2-841d-3f096c3adb18\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-whc4s" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.634258 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.635092 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/523e6f72-37ee-4c32-b3e0-98405775373f-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-29mhg\" (UID: \"523e6f72-37ee-4c32-b3e0-98405775373f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-29mhg" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.640889 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hqjh4" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.648239 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-tc862" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.656178 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hvkf\" (UniqueName: \"kubernetes.io/projected/ee393336-510d-40f3-a382-06d7f839750f-kube-api-access-8hvkf\") pod \"packageserver-d55dfcdfc-r8xzr\" (UID: \"ee393336-510d-40f3-a382-06d7f839750f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r8xzr" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.663444 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jczvb" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.666235 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:29 crc kubenswrapper[4787]: E1001 09:38:29.666860 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:30.166839061 +0000 UTC m=+142.281983218 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.676571 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpsvj\" (UniqueName: \"kubernetes.io/projected/11d20ebc-a959-43fa-84eb-bdc70a3f38f4-kube-api-access-dpsvj\") pod \"catalog-operator-68c6474976-rv6xx\" (UID: \"11d20ebc-a959-43fa-84eb-bdc70a3f38f4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rv6xx" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.693336 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7f2hj\" (UniqueName: \"kubernetes.io/projected/2ec1df2f-eadd-4385-97b4-664101fb9d51-kube-api-access-7f2hj\") pod \"console-f9d7485db-gqvt9\" (UID: \"2ec1df2f-eadd-4385-97b4-664101fb9d51\") " pod="openshift-console/console-f9d7485db-gqvt9" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.702270 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-t2mbp"] Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.713497 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dl48\" (UniqueName: \"kubernetes.io/projected/a2b762fa-551a-40c6-acd8-45e6649949d6-kube-api-access-6dl48\") pod \"ingress-canary-sgxp6\" (UID: \"a2b762fa-551a-40c6-acd8-45e6649949d6\") " pod="openshift-ingress-canary/ingress-canary-sgxp6" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.721405 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gjtg8" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.726889 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e9338499-2f07-4087-b75f-bf85d8a2a794-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-86dvz\" (UID: \"e9338499-2f07-4087-b75f-bf85d8a2a794\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-86dvz" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.732598 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dscrd" Oct 01 09:38:29 crc kubenswrapper[4787]: W1001 09:38:29.750558 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod15f1cb19_94ee_49b2_a837_b7bdaf484387.slice/crio-b937b5092e25919724be22c673675f209e2387e7cec9124fef9e43d0b668ac17 WatchSource:0}: Error finding container b937b5092e25919724be22c673675f209e2387e7cec9124fef9e43d0b668ac17: Status 404 returned error can't find the container with id b937b5092e25919724be22c673675f209e2387e7cec9124fef9e43d0b668ac17 Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.756916 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtwkc\" (UniqueName: \"kubernetes.io/projected/e697b985-6356-4777-8ed2-06f9797a449c-kube-api-access-jtwkc\") pod \"machine-config-server-ns9tt\" (UID: \"e697b985-6356-4777-8ed2-06f9797a449c\") " pod="openshift-machine-config-operator/machine-config-server-ns9tt" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.769975 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:29 crc kubenswrapper[4787]: E1001 09:38:29.770750 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:30.270727011 +0000 UTC m=+142.385871168 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.772920 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dm6j\" (UniqueName: \"kubernetes.io/projected/8de4a645-2945-4261-b9be-ee3336e818df-kube-api-access-6dm6j\") pod \"collect-profiles-29321850-kw99x\" (UID: \"8de4a645-2945-4261-b9be-ee3336e818df\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-kw99x" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.775263 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-lld4x" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.795137 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzljp\" (UniqueName: \"kubernetes.io/projected/5d7d263b-bc04-4069-81a5-d5cd1151d805-kube-api-access-gzljp\") pod \"marketplace-operator-79b997595-cf76m\" (UID: \"5d7d263b-bc04-4069-81a5-d5cd1151d805\") " pod="openshift-marketplace/marketplace-operator-79b997595-cf76m" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.807437 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qvdw4" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.815035 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxgz5\" (UniqueName: \"kubernetes.io/projected/aca7d6cc-eca3-4ffb-84d9-254616d0c037-kube-api-access-gxgz5\") pod \"dns-operator-744455d44c-fsblz\" (UID: \"aca7d6cc-eca3-4ffb-84d9-254616d0c037\") " pod="openshift-dns-operator/dns-operator-744455d44c-fsblz" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.818833 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-fsblz" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.828937 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-46vzw"] Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.836293 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fzd5j"] Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.844169 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-29mhg" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.844639 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rv6xx" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.845054 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/54dbcab2-1757-4f7d-b59c-128e257b5660-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9x4nl\" (UID: \"54dbcab2-1757-4f7d-b59c-128e257b5660\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9x4nl" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.854456 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-whc4s" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.863116 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-86dvz" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.866489 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7nkk\" (UniqueName: \"kubernetes.io/projected/151f124f-e071-4884-aeb4-217064807fa8-kube-api-access-v7nkk\") pod \"csi-hostpathplugin-ncnkh\" (UID: \"151f124f-e071-4884-aeb4-217064807fa8\") " pod="hostpath-provisioner/csi-hostpathplugin-ncnkh" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.871556 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.871908 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r8xzr" Oct 01 09:38:29 crc kubenswrapper[4787]: E1001 09:38:29.872020 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:30.372004863 +0000 UTC m=+142.487149020 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.874870 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4xvq\" (UniqueName: \"kubernetes.io/projected/e6b988c0-936d-4b49-80a2-015ae390f262-kube-api-access-t4xvq\") pod \"control-plane-machine-set-operator-78cbb6b69f-zssrz\" (UID: \"e6b988c0-936d-4b49-80a2-015ae390f262\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zssrz" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.887058 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zssrz" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.892660 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hr4xq\" (UniqueName: \"kubernetes.io/projected/75578ed4-3bb2-4ed7-bac1-d28780be4d67-kube-api-access-hr4xq\") pod \"multus-admission-controller-857f4d67dd-n6959\" (UID: \"75578ed4-3bb2-4ed7-bac1-d28780be4d67\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-n6959" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.896399 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-kw99x" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.913822 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-gqvt9" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.916895 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwmhz\" (UniqueName: \"kubernetes.io/projected/0ed372ed-3529-4e31-b124-d66f292089d0-kube-api-access-xwmhz\") pod \"machine-approver-56656f9798-pzdzn\" (UID: \"0ed372ed-3529-4e31-b124-d66f292089d0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pzdzn" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.926036 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfxzm\" (UniqueName: \"kubernetes.io/projected/13cafd2b-3d1e-4d8b-8751-a583af048423-kube-api-access-qfxzm\") pod \"package-server-manager-789f6589d5-blwqt\" (UID: \"13cafd2b-3d1e-4d8b-8751-a583af048423\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-blwqt" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.951432 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cf76m" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.955599 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4mj7\" (UniqueName: \"kubernetes.io/projected/59fc7ef6-b52d-41e5-8c82-9ce4ed5d5053-kube-api-access-d4mj7\") pod \"service-ca-9c57cc56f-tn48z\" (UID: \"59fc7ef6-b52d-41e5-8c82-9ce4ed5d5053\") " pod="openshift-service-ca/service-ca-9c57cc56f-tn48z" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.961779 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-ns9tt" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.974986 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0f1d5b79-c7bd-4c71-aa2c-7e407c4b7adb-bound-sa-token\") pod \"ingress-operator-5b745b69d9-n47zx\" (UID: \"0f1d5b79-c7bd-4c71-aa2c-7e407c4b7adb\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n47zx" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.977490 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-sgxp6" Oct 01 09:38:29 crc kubenswrapper[4787]: I1001 09:38:29.978280 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:29 crc kubenswrapper[4787]: E1001 09:38:29.978731 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:30.478713725 +0000 UTC m=+142.593857882 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:30 crc kubenswrapper[4787]: I1001 09:38:29.999508 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-px5h8\" (UniqueName: \"kubernetes.io/projected/9f2a2142-5430-4edb-8f20-b199d0425466-kube-api-access-px5h8\") pod \"dns-default-9lvh8\" (UID: \"9f2a2142-5430-4edb-8f20-b199d0425466\") " pod="openshift-dns/dns-default-9lvh8" Oct 01 09:38:30 crc kubenswrapper[4787]: I1001 09:38:30.000197 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-ncnkh" Oct 01 09:38:30 crc kubenswrapper[4787]: I1001 09:38:30.040657 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hflzd\" (UniqueName: \"kubernetes.io/projected/d438f633-0698-4c39-9c5a-71a57a6b3fbc-kube-api-access-hflzd\") pod \"console-operator-58897d9998-hqx7l\" (UID: \"d438f633-0698-4c39-9c5a-71a57a6b3fbc\") " pod="openshift-console-operator/console-operator-58897d9998-hqx7l" Oct 01 09:38:30 crc kubenswrapper[4787]: I1001 09:38:30.059391 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9lfw\" (UniqueName: \"kubernetes.io/projected/0f1d5b79-c7bd-4c71-aa2c-7e407c4b7adb-kube-api-access-h9lfw\") pod \"ingress-operator-5b745b69d9-n47zx\" (UID: \"0f1d5b79-c7bd-4c71-aa2c-7e407c4b7adb\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n47zx" Oct 01 09:38:30 crc kubenswrapper[4787]: I1001 09:38:30.067597 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pzdzn" Oct 01 09:38:30 crc kubenswrapper[4787]: I1001 09:38:30.068767 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49rkp\" (UniqueName: \"kubernetes.io/projected/5c56c7aa-b794-474b-910b-11ecf8481009-kube-api-access-49rkp\") pod \"etcd-operator-b45778765-sgphp\" (UID: \"5c56c7aa-b794-474b-910b-11ecf8481009\") " pod="openshift-etcd-operator/etcd-operator-b45778765-sgphp" Oct 01 09:38:30 crc kubenswrapper[4787]: I1001 09:38:30.085057 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:30 crc kubenswrapper[4787]: E1001 09:38:30.088373 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:30.588350111 +0000 UTC m=+142.703494268 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:30 crc kubenswrapper[4787]: I1001 09:38:30.096617 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-hqx7l" Oct 01 09:38:30 crc kubenswrapper[4787]: I1001 09:38:30.131274 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9x4nl" Oct 01 09:38:30 crc kubenswrapper[4787]: I1001 09:38:30.135514 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n47zx" Oct 01 09:38:30 crc kubenswrapper[4787]: I1001 09:38:30.181302 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-n6959" Oct 01 09:38:30 crc kubenswrapper[4787]: I1001 09:38:30.187617 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:30 crc kubenswrapper[4787]: E1001 09:38:30.188009 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:30.687993492 +0000 UTC m=+142.803137649 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:30 crc kubenswrapper[4787]: I1001 09:38:30.202713 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-tn48z" Oct 01 09:38:30 crc kubenswrapper[4787]: I1001 09:38:30.220869 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-blwqt" Oct 01 09:38:30 crc kubenswrapper[4787]: I1001 09:38:30.232762 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-sgphp" Oct 01 09:38:30 crc kubenswrapper[4787]: I1001 09:38:30.233191 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jczvb"] Oct 01 09:38:30 crc kubenswrapper[4787]: I1001 09:38:30.249837 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-hwftw"] Oct 01 09:38:30 crc kubenswrapper[4787]: I1001 09:38:30.283425 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-9lvh8" Oct 01 09:38:30 crc kubenswrapper[4787]: I1001 09:38:30.289578 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:30 crc kubenswrapper[4787]: E1001 09:38:30.290166 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:30.790145597 +0000 UTC m=+142.905289754 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:30 crc kubenswrapper[4787]: I1001 09:38:30.303401 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-tc862"] Oct 01 09:38:30 crc kubenswrapper[4787]: I1001 09:38:30.318708 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-46vzw" event={"ID":"b0ea3010-e975-4d1e-9fa4-385489660daa","Type":"ContainerStarted","Data":"8144c9553b5003c8637607d9db57ca1cb788d8dbc820cdd3135c02803c9a02b7"} Oct 01 09:38:30 crc kubenswrapper[4787]: I1001 09:38:30.326595 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhp4" event={"ID":"160b7ce2-3861-4c15-af76-2b0d41ec8301","Type":"ContainerStarted","Data":"ae28d6d7f6b1c496e94ded4c8b979204ba989d567c624eb4a70383c63be7a52b"} Oct 01 09:38:30 crc kubenswrapper[4787]: I1001 09:38:30.332259 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t2mbp" event={"ID":"15f1cb19-94ee-49b2-a837-b7bdaf484387","Type":"ContainerStarted","Data":"83a144991dd19a3d40f1fcaa174a5e4e7cb939445cac1ffe651f330392bb28b9"} Oct 01 09:38:30 crc kubenswrapper[4787]: I1001 09:38:30.332323 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t2mbp" event={"ID":"15f1cb19-94ee-49b2-a837-b7bdaf484387","Type":"ContainerStarted","Data":"b937b5092e25919724be22c673675f209e2387e7cec9124fef9e43d0b668ac17"} Oct 01 09:38:30 crc kubenswrapper[4787]: I1001 09:38:30.336413 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tvvw6" event={"ID":"1c07acd7-0397-494c-ab52-83bcdefdf2ac","Type":"ContainerStarted","Data":"5d727064a4d3f6cb260877a7412be1bb2b75a254fe942d809ad1814190a31333"} Oct 01 09:38:30 crc kubenswrapper[4787]: I1001 09:38:30.336460 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tvvw6" event={"ID":"1c07acd7-0397-494c-ab52-83bcdefdf2ac","Type":"ContainerStarted","Data":"23454696854b61478f0cb374e92b1c19556296a6f69732ef2292891bd4d7fe6b"} Oct 01 09:38:30 crc kubenswrapper[4787]: I1001 09:38:30.337860 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fzd5j" event={"ID":"652585ff-b553-4326-add1-ba20d0dd0a11","Type":"ContainerStarted","Data":"8eb222144145f25e2643088e7b78fe2dab025244aa0da89dc5ba2d4dc6d36868"} Oct 01 09:38:30 crc kubenswrapper[4787]: I1001 09:38:30.340970 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-7p4xb" event={"ID":"52a18549-2c1c-4f28-bfa1-6fb14e0690e1","Type":"ContainerStarted","Data":"fca50372bcd94b83d28d343ca9aded4b37e420a220f5254a8cbb78cef97f6614"} Oct 01 09:38:30 crc kubenswrapper[4787]: I1001 09:38:30.353060 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-djtn2" Oct 01 09:38:30 crc kubenswrapper[4787]: I1001 09:38:30.390366 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:30 crc kubenswrapper[4787]: E1001 09:38:30.390867 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:30.890848214 +0000 UTC m=+143.005992371 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:30 crc kubenswrapper[4787]: I1001 09:38:30.416539 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-x649j" Oct 01 09:38:30 crc kubenswrapper[4787]: I1001 09:38:30.494801 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:30 crc kubenswrapper[4787]: E1001 09:38:30.497780 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:30.997758962 +0000 UTC m=+143.112903119 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:30 crc kubenswrapper[4787]: I1001 09:38:30.505855 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-hqjh4"] Oct 01 09:38:30 crc kubenswrapper[4787]: I1001 09:38:30.597460 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:30 crc kubenswrapper[4787]: E1001 09:38:30.597616 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:31.097593537 +0000 UTC m=+143.212737694 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:30 crc kubenswrapper[4787]: I1001 09:38:30.597721 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:30 crc kubenswrapper[4787]: E1001 09:38:30.598039 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:31.098032968 +0000 UTC m=+143.213177125 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:30 crc kubenswrapper[4787]: I1001 09:38:30.667388 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-tzsx8"] Oct 01 09:38:30 crc kubenswrapper[4787]: I1001 09:38:30.706816 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:30 crc kubenswrapper[4787]: E1001 09:38:30.707366 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:31.207343978 +0000 UTC m=+143.322488125 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:30 crc kubenswrapper[4787]: W1001 09:38:30.781300 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode15a5f43_d273_42fb_b8d2_ba421fb1ab64.slice/crio-dd63ec3766358af7ae1d2dd9ed89e8aa8d0ba3b53cea2f2be47cc792f0347389 WatchSource:0}: Error finding container dd63ec3766358af7ae1d2dd9ed89e8aa8d0ba3b53cea2f2be47cc792f0347389: Status 404 returned error can't find the container with id dd63ec3766358af7ae1d2dd9ed89e8aa8d0ba3b53cea2f2be47cc792f0347389 Oct 01 09:38:30 crc kubenswrapper[4787]: I1001 09:38:30.809200 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:30 crc kubenswrapper[4787]: E1001 09:38:30.809548 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:31.309536074 +0000 UTC m=+143.424680231 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:30 crc kubenswrapper[4787]: I1001 09:38:30.867305 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-djtn2" podStartSLOduration=122.867286152 podStartE2EDuration="2m2.867286152s" podCreationTimestamp="2025-10-01 09:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:30.833016904 +0000 UTC m=+142.948161081" watchObservedRunningTime="2025-10-01 09:38:30.867286152 +0000 UTC m=+142.982430309" Oct 01 09:38:30 crc kubenswrapper[4787]: I1001 09:38:30.909925 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:30 crc kubenswrapper[4787]: E1001 09:38:30.910414 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:31.410393626 +0000 UTC m=+143.525537773 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.025207 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:31 crc kubenswrapper[4787]: E1001 09:38:31.026073 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:31.526049326 +0000 UTC m=+143.641193483 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.057283 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhp4" podStartSLOduration=123.057266085 podStartE2EDuration="2m3.057266085s" podCreationTimestamp="2025-10-01 09:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:31.023012678 +0000 UTC m=+143.138156855" watchObservedRunningTime="2025-10-01 09:38:31.057266085 +0000 UTC m=+143.172410242" Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.107515 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-gjtg8"] Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.113024 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-7p4xb" Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.118472 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rv6xx"] Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.144462 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:31 crc kubenswrapper[4787]: E1001 09:38:31.145946 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:31.645919595 +0000 UTC m=+143.761063752 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.224640 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dscrd"] Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.253103 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:31 crc kubenswrapper[4787]: E1001 09:38:31.253486 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:31.753471378 +0000 UTC m=+143.868615535 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.355178 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:31 crc kubenswrapper[4787]: E1001 09:38:31.355883 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:31.855852579 +0000 UTC m=+143.970996736 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.358045 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jczvb" event={"ID":"aa998110-e80e-4bc1-b1b2-f9806030ba26","Type":"ContainerStarted","Data":"c75e4987ae815343d24eb17ab27901c6ca64da22a5d52557e772cfacca4c3013"} Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.358124 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jczvb" event={"ID":"aa998110-e80e-4bc1-b1b2-f9806030ba26","Type":"ContainerStarted","Data":"a3dda2bc81a441d619f7dedf21a6e77166f3a6fbd923d7353523215d3ceedf83"} Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.362538 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tvvw6" event={"ID":"1c07acd7-0397-494c-ab52-83bcdefdf2ac","Type":"ContainerStarted","Data":"35041af8e99636c044c26d7c9ee6eab04516211c11abd960aeb6b5545ce9495b"} Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.368060 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hqjh4" event={"ID":"e15a5f43-d273-42fb-b8d2-ba421fb1ab64","Type":"ContainerStarted","Data":"dd63ec3766358af7ae1d2dd9ed89e8aa8d0ba3b53cea2f2be47cc792f0347389"} Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.374651 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fzd5j" event={"ID":"652585ff-b553-4326-add1-ba20d0dd0a11","Type":"ContainerStarted","Data":"a9ce27ff9cdf736c28c2698f5e941825b98eb2529d5dacfcbb7f81f00b24ab69"} Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.375680 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-x649j" podStartSLOduration=123.375658516 podStartE2EDuration="2m3.375658516s" podCreationTimestamp="2025-10-01 09:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:31.372741611 +0000 UTC m=+143.487885778" watchObservedRunningTime="2025-10-01 09:38:31.375658516 +0000 UTC m=+143.490802673" Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.383646 4787 generic.go:334] "Generic (PLEG): container finished" podID="b0ea3010-e975-4d1e-9fa4-385489660daa" containerID="c8fa8afe8158bbed5ca7f68be414981945a4f0b98b474433090616bc46765cdb" exitCode=0 Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.383761 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-46vzw" event={"ID":"b0ea3010-e975-4d1e-9fa4-385489660daa","Type":"ContainerDied","Data":"c8fa8afe8158bbed5ca7f68be414981945a4f0b98b474433090616bc46765cdb"} Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.385672 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" event={"ID":"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e","Type":"ContainerStarted","Data":"192a63e14bec77d4764e90ac496ddeefd96b9dff73c66657597f93f3c37901bb"} Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.391747 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-tc862" event={"ID":"b160cc9e-253c-4aba-951a-21fd20ff52f4","Type":"ContainerStarted","Data":"749a81e1b804eb4dfdf682e12e328e3bd99ceb7dd76868f3b38b321c77aac0b9"} Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.391802 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-tc862" event={"ID":"b160cc9e-253c-4aba-951a-21fd20ff52f4","Type":"ContainerStarted","Data":"d90162b134a210e5d8b88780a094727d5b505af582ee6a9c51ba654716f870df"} Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.392344 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-tc862" Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.394511 4787 patch_prober.go:28] interesting pod/downloads-7954f5f757-tc862 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.394555 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-tc862" podUID="b160cc9e-253c-4aba-951a-21fd20ff52f4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.394754 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pzdzn" event={"ID":"0ed372ed-3529-4e31-b124-d66f292089d0","Type":"ContainerStarted","Data":"87cb5e94aa4381f82f9f84edfcbe2e217311c873b9158c6cd0e85105e9ffa944"} Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.396569 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t2mbp" event={"ID":"15f1cb19-94ee-49b2-a837-b7bdaf484387","Type":"ContainerStarted","Data":"3e76dbe2a1b8d1c036323a7b67af17b49972dc485db888bf6a6244ce750edf4f"} Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.397616 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tzsx8" event={"ID":"fa4ba99a-cfa6-4e3b-bd6f-ce18ea3fefe8","Type":"ContainerStarted","Data":"12cec42cbe92eb1c02cfba10978ce3b44e6abeabec2724ea3efbf86d2dfc39e3"} Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.408460 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-ns9tt" event={"ID":"e697b985-6356-4777-8ed2-06f9797a449c","Type":"ContainerStarted","Data":"402aee35fb5d21f46c47f7e2f1fa00cd6c56a12bf103f90b5e4967de37ca6d04"} Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.408527 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-ns9tt" event={"ID":"e697b985-6356-4777-8ed2-06f9797a449c","Type":"ContainerStarted","Data":"c4d68b4fa886ff8bf87b20f5c29eae7486e85efb3d0818459d2411258af1696d"} Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.414877 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-cdc2f" podStartSLOduration=124.414855839 podStartE2EDuration="2m4.414855839s" podCreationTimestamp="2025-10-01 09:36:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:31.411520303 +0000 UTC m=+143.526664460" watchObservedRunningTime="2025-10-01 09:38:31.414855839 +0000 UTC m=+143.529999996" Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.458734 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-4g5xc" podStartSLOduration=123.458718332 podStartE2EDuration="2m3.458718332s" podCreationTimestamp="2025-10-01 09:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:31.456796943 +0000 UTC m=+143.571941120" watchObservedRunningTime="2025-10-01 09:38:31.458718332 +0000 UTC m=+143.573862489" Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.460657 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:31 crc kubenswrapper[4787]: E1001 09:38:31.469944 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:31.969926199 +0000 UTC m=+144.085070346 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.561781 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:31 crc kubenswrapper[4787]: E1001 09:38:31.562197 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:32.06216581 +0000 UTC m=+144.177309967 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.562409 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:31 crc kubenswrapper[4787]: E1001 09:38:31.563032 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:32.063012892 +0000 UTC m=+144.178157059 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.667810 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:31 crc kubenswrapper[4787]: E1001 09:38:31.668238 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:32.168198385 +0000 UTC m=+144.283342542 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.668440 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:31 crc kubenswrapper[4787]: E1001 09:38:31.668919 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:32.168901652 +0000 UTC m=+144.284045819 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.685577 4787 patch_prober.go:28] interesting pod/router-default-5444994796-7p4xb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 09:38:31 crc kubenswrapper[4787]: [-]has-synced failed: reason withheld Oct 01 09:38:31 crc kubenswrapper[4787]: [+]process-running ok Oct 01 09:38:31 crc kubenswrapper[4787]: healthz check failed Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.685637 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7p4xb" podUID="52a18549-2c1c-4f28-bfa1-6fb14e0690e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 09:38:31 crc kubenswrapper[4787]: W1001 09:38:31.686598 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod11d20ebc_a959_43fa_84eb_bdc70a3f38f4.slice/crio-c852bbc215485f7ba987389a28987929717983ce4967549d0bae1fcdd484d191 WatchSource:0}: Error finding container c852bbc215485f7ba987389a28987929717983ce4967549d0bae1fcdd484d191: Status 404 returned error can't find the container with id c852bbc215485f7ba987389a28987929717983ce4967549d0bae1fcdd484d191 Oct 01 09:38:31 crc kubenswrapper[4787]: W1001 09:38:31.724877 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod57a3e335_0a9a_4fb5_b234_e0017b775ecd.slice/crio-922b13dcb6d5cc081e8321cf354a93aeeef51dd4ffcd03b930bc82c4093904f0 WatchSource:0}: Error finding container 922b13dcb6d5cc081e8321cf354a93aeeef51dd4ffcd03b930bc82c4093904f0: Status 404 returned error can't find the container with id 922b13dcb6d5cc081e8321cf354a93aeeef51dd4ffcd03b930bc82c4093904f0 Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.770460 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:31 crc kubenswrapper[4787]: E1001 09:38:31.770826 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:32.270809962 +0000 UTC m=+144.385954119 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.841107 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-7p4xb" podStartSLOduration=123.84106112 podStartE2EDuration="2m3.84106112s" podCreationTimestamp="2025-10-01 09:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:31.810068957 +0000 UTC m=+143.925213134" watchObservedRunningTime="2025-10-01 09:38:31.84106112 +0000 UTC m=+143.956205267" Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.872406 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:31 crc kubenswrapper[4787]: E1001 09:38:31.872796 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:32.372784992 +0000 UTC m=+144.487929149 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:31 crc kubenswrapper[4787]: I1001 09:38:31.973833 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:31 crc kubenswrapper[4787]: E1001 09:38:31.974188 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:32.474170187 +0000 UTC m=+144.589314334 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.075067 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:32 crc kubenswrapper[4787]: E1001 09:38:32.075444 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:32.57542637 +0000 UTC m=+144.690570527 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.080402 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-tc862" podStartSLOduration=124.080387127 podStartE2EDuration="2m4.080387127s" podCreationTimestamp="2025-10-01 09:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:32.026913898 +0000 UTC m=+144.142058065" watchObservedRunningTime="2025-10-01 09:38:32.080387127 +0000 UTC m=+144.195531284" Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.080481 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jczvb" podStartSLOduration=124.080477639 podStartE2EDuration="2m4.080477639s" podCreationTimestamp="2025-10-01 09:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:32.078468278 +0000 UTC m=+144.193612435" watchObservedRunningTime="2025-10-01 09:38:32.080477639 +0000 UTC m=+144.195621796" Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.108356 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t2mbp" podStartSLOduration=124.108331372 podStartE2EDuration="2m4.108331372s" podCreationTimestamp="2025-10-01 09:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:32.105734026 +0000 UTC m=+144.220878193" watchObservedRunningTime="2025-10-01 09:38:32.108331372 +0000 UTC m=+144.223475529" Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.170337 4787 patch_prober.go:28] interesting pod/router-default-5444994796-7p4xb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 09:38:32 crc kubenswrapper[4787]: [-]has-synced failed: reason withheld Oct 01 09:38:32 crc kubenswrapper[4787]: [+]process-running ok Oct 01 09:38:32 crc kubenswrapper[4787]: healthz check failed Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.170392 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7p4xb" podUID="52a18549-2c1c-4f28-bfa1-6fb14e0690e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.176458 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:32 crc kubenswrapper[4787]: E1001 09:38:32.177931 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:32.677913294 +0000 UTC m=+144.793057451 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.233735 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fzd5j" podStartSLOduration=124.233698332 podStartE2EDuration="2m4.233698332s" podCreationTimestamp="2025-10-01 09:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:32.229498605 +0000 UTC m=+144.344642762" watchObservedRunningTime="2025-10-01 09:38:32.233698332 +0000 UTC m=+144.348842479" Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.276884 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-ns9tt" podStartSLOduration=5.276853267 podStartE2EDuration="5.276853267s" podCreationTimestamp="2025-10-01 09:38:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:32.271256894 +0000 UTC m=+144.386401051" watchObservedRunningTime="2025-10-01 09:38:32.276853267 +0000 UTC m=+144.391997424" Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.290700 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:32 crc kubenswrapper[4787]: E1001 09:38:32.291320 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:32.791304207 +0000 UTC m=+144.906448364 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.298976 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tvvw6" podStartSLOduration=124.298815219 podStartE2EDuration="2m4.298815219s" podCreationTimestamp="2025-10-01 09:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:32.297375972 +0000 UTC m=+144.412520159" watchObservedRunningTime="2025-10-01 09:38:32.298815219 +0000 UTC m=+144.413959376" Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.369502 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r8xzr"] Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.391482 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.391756 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-fsblz"] Oct 01 09:38:32 crc kubenswrapper[4787]: E1001 09:38:32.391954 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:32.891937713 +0000 UTC m=+145.007081870 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:32 crc kubenswrapper[4787]: W1001 09:38:32.416755 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaca7d6cc_eca3_4ffb_84d9_254616d0c037.slice/crio-c32e8d33d79150cf8ac08b055ebe31f092ed3ad74dba8d40eee5e0ea0e9c9934 WatchSource:0}: Error finding container c32e8d33d79150cf8ac08b055ebe31f092ed3ad74dba8d40eee5e0ea0e9c9934: Status 404 returned error can't find the container with id c32e8d33d79150cf8ac08b055ebe31f092ed3ad74dba8d40eee5e0ea0e9c9934 Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.428421 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rv6xx" event={"ID":"11d20ebc-a959-43fa-84eb-bdc70a3f38f4","Type":"ContainerStarted","Data":"a36aa1d1180f137548aa9cbf17b72d7b71f0c0987680f20b75d553c41a1fdf9c"} Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.428466 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rv6xx" event={"ID":"11d20ebc-a959-43fa-84eb-bdc70a3f38f4","Type":"ContainerStarted","Data":"c852bbc215485f7ba987389a28987929717983ce4967549d0bae1fcdd484d191"} Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.429511 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rv6xx" Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.435386 4787 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-rv6xx container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" start-of-body= Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.435453 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rv6xx" podUID="11d20ebc-a959-43fa-84eb-bdc70a3f38f4" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.442702 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gjtg8" event={"ID":"82a8aff4-187a-464a-b375-5c20e576164d","Type":"ContainerStarted","Data":"ea28797da707f728cc36dac165152b5946b9df0f761c2e64bf5568447714c820"} Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.442742 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gjtg8" event={"ID":"82a8aff4-187a-464a-b375-5c20e576164d","Type":"ContainerStarted","Data":"ee62870606e89269038b72897d89ee53163dec95514af7e38cabd4aa541fcea4"} Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.470620 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rv6xx" podStartSLOduration=124.470597257 podStartE2EDuration="2m4.470597257s" podCreationTimestamp="2025-10-01 09:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:32.460302583 +0000 UTC m=+144.575446740" watchObservedRunningTime="2025-10-01 09:38:32.470597257 +0000 UTC m=+144.585741414" Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.484936 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" event={"ID":"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e","Type":"ContainerStarted","Data":"d68100bc22c5dd781e9f555c26fcc97f3da43c3cc79dc00c997b00804c9dd1e1"} Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.485482 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.490178 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dscrd" event={"ID":"57a3e335-0a9a-4fb5-b234-e0017b775ecd","Type":"ContainerStarted","Data":"33b49d5a77458eb227705dfbae3bbe0b0fb4f3bd57c5ad7e6a6333d6eab34370"} Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.490229 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dscrd" event={"ID":"57a3e335-0a9a-4fb5-b234-e0017b775ecd","Type":"ContainerStarted","Data":"922b13dcb6d5cc081e8321cf354a93aeeef51dd4ffcd03b930bc82c4093904f0"} Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.493018 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.493972 4787 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-hwftw container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.10:6443/healthz\": dial tcp 10.217.0.10:6443: connect: connection refused" start-of-body= Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.494037 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" podUID="7ef82a9b-8896-4ff4-ae93-91ea7ed2885e" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.10:6443/healthz\": dial tcp 10.217.0.10:6443: connect: connection refused" Oct 01 09:38:32 crc kubenswrapper[4787]: E1001 09:38:32.494444 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:32.994431107 +0000 UTC m=+145.109575264 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.505106 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qvdw4"] Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.514785 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pzdzn" event={"ID":"0ed372ed-3529-4e31-b124-d66f292089d0","Type":"ContainerStarted","Data":"59fea6e9c26d5807d1d1c77c97b86f33950a00bcb6feb725f2512cef3eb93a71"} Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.553873 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" podStartSLOduration=125.553854728 podStartE2EDuration="2m5.553854728s" podCreationTimestamp="2025-10-01 09:36:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:32.52310211 +0000 UTC m=+144.638246277" watchObservedRunningTime="2025-10-01 09:38:32.553854728 +0000 UTC m=+144.668998875" Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.569553 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dscrd" podStartSLOduration=124.569516019 podStartE2EDuration="2m4.569516019s" podCreationTimestamp="2025-10-01 09:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:32.553118999 +0000 UTC m=+144.668263156" watchObservedRunningTime="2025-10-01 09:38:32.569516019 +0000 UTC m=+144.684660176" Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.569820 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-whc4s"] Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.590664 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tzsx8" event={"ID":"fa4ba99a-cfa6-4e3b-bd6f-ce18ea3fefe8","Type":"ContainerStarted","Data":"9918edc2a80e5bd97c0b3380af2abd30cb29e064d133f347ad0411022ed79697"} Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.602803 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:32 crc kubenswrapper[4787]: E1001 09:38:32.604279 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:33.104253648 +0000 UTC m=+145.219397815 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.685967 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-lld4x"] Oct 01 09:38:32 crc kubenswrapper[4787]: W1001 09:38:32.689496 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod556c69e8_1179_4de2_841d_3f096c3adb18.slice/crio-865605b6fb543b69e55d98063b717689fe8b9681898ab76270069c6d8eb863e2 WatchSource:0}: Error finding container 865605b6fb543b69e55d98063b717689fe8b9681898ab76270069c6d8eb863e2: Status 404 returned error can't find the container with id 865605b6fb543b69e55d98063b717689fe8b9681898ab76270069c6d8eb863e2 Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.693230 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-46vzw" event={"ID":"b0ea3010-e975-4d1e-9fa4-385489660daa","Type":"ContainerStarted","Data":"8551d32c2097fddf905705111d95e9ae719ef4365bc29ec80b42d2eccba08af3"} Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.714906 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-blwqt"] Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.720929 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r8xzr" event={"ID":"ee393336-510d-40f3-a382-06d7f839750f","Type":"ContainerStarted","Data":"0a0ec8bf354c28c8244053770f5c55b34b6207e120dd543ebfa670039bc166e3"} Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.737347 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:32 crc kubenswrapper[4787]: E1001 09:38:32.737752 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:33.237739016 +0000 UTC m=+145.352883183 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.745497 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-29mhg"] Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.753537 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-86dvz"] Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.783598 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hqjh4" event={"ID":"e15a5f43-d273-42fb-b8d2-ba421fb1ab64","Type":"ContainerStarted","Data":"79f158f94f57cb897d92270be5be8c994b51fabe9a0df1798b3dc7c0fc741e69"} Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.793545 4787 patch_prober.go:28] interesting pod/downloads-7954f5f757-tc862 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.793869 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-tc862" podUID="b160cc9e-253c-4aba-951a-21fd20ff52f4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Oct 01 09:38:32 crc kubenswrapper[4787]: W1001 09:38:32.816989 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod523e6f72_37ee_4c32_b3e0_98405775373f.slice/crio-b46fe9a7ad75c2f69de3672d9eecfd10e1218751ebe134704d96685b5a4f966d WatchSource:0}: Error finding container b46fe9a7ad75c2f69de3672d9eecfd10e1218751ebe134704d96685b5a4f966d: Status 404 returned error can't find the container with id b46fe9a7ad75c2f69de3672d9eecfd10e1218751ebe134704d96685b5a4f966d Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.838322 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:32 crc kubenswrapper[4787]: E1001 09:38:32.838776 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:33.338754901 +0000 UTC m=+145.453899058 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.855137 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hqjh4" podStartSLOduration=124.85510919 podStartE2EDuration="2m4.85510919s" podCreationTimestamp="2025-10-01 09:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:32.826724213 +0000 UTC m=+144.941868380" watchObservedRunningTime="2025-10-01 09:38:32.85510919 +0000 UTC m=+144.970253347" Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.855500 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-gqvt9"] Oct 01 09:38:32 crc kubenswrapper[4787]: W1001 09:38:32.867386 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode9338499_2f07_4087_b75f_bf85d8a2a794.slice/crio-50808a82f6cfab13a3b3041aa2d676d9396112da1d8ccc8623cccd14c61b1b74 WatchSource:0}: Error finding container 50808a82f6cfab13a3b3041aa2d676d9396112da1d8ccc8623cccd14c61b1b74: Status 404 returned error can't find the container with id 50808a82f6cfab13a3b3041aa2d676d9396112da1d8ccc8623cccd14c61b1b74 Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.901649 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-hqx7l"] Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.929988 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321850-kw99x"] Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.943492 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:32 crc kubenswrapper[4787]: E1001 09:38:32.950287 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:33.450263715 +0000 UTC m=+145.565408072 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:32 crc kubenswrapper[4787]: I1001 09:38:32.957178 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zssrz"] Oct 01 09:38:33 crc kubenswrapper[4787]: W1001 09:38:33.001244 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2ec1df2f_eadd_4385_97b4_664101fb9d51.slice/crio-7fb44713d867d5f5b7e61272f9947cc5ddbb9c637b20236ca48d6ba01ac38bb3 WatchSource:0}: Error finding container 7fb44713d867d5f5b7e61272f9947cc5ddbb9c637b20236ca48d6ba01ac38bb3: Status 404 returned error can't find the container with id 7fb44713d867d5f5b7e61272f9947cc5ddbb9c637b20236ca48d6ba01ac38bb3 Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.041509 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-n6959"] Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.050387 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:33 crc kubenswrapper[4787]: E1001 09:38:33.052248 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:33.552220206 +0000 UTC m=+145.667364363 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.061807 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-ncnkh"] Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.061872 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-tn48z"] Oct 01 09:38:33 crc kubenswrapper[4787]: W1001 09:38:33.062504 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8de4a645_2945_4261_b9be_ee3336e818df.slice/crio-4bd1a34abf91ed08f32088d26f36a1ce8ad1e6a1b960b35b45aeda0de9a12b9c WatchSource:0}: Error finding container 4bd1a34abf91ed08f32088d26f36a1ce8ad1e6a1b960b35b45aeda0de9a12b9c: Status 404 returned error can't find the container with id 4bd1a34abf91ed08f32088d26f36a1ce8ad1e6a1b960b35b45aeda0de9a12b9c Oct 01 09:38:33 crc kubenswrapper[4787]: W1001 09:38:33.089333 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode6b988c0_936d_4b49_80a2_015ae390f262.slice/crio-52764d4206fe707c89308dece827cde4bcb8906f36ced74544221b6a785936c1 WatchSource:0}: Error finding container 52764d4206fe707c89308dece827cde4bcb8906f36ced74544221b6a785936c1: Status 404 returned error can't find the container with id 52764d4206fe707c89308dece827cde4bcb8906f36ced74544221b6a785936c1 Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.090653 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9x4nl"] Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.137453 4787 patch_prober.go:28] interesting pod/router-default-5444994796-7p4xb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 09:38:33 crc kubenswrapper[4787]: [-]has-synced failed: reason withheld Oct 01 09:38:33 crc kubenswrapper[4787]: [+]process-running ok Oct 01 09:38:33 crc kubenswrapper[4787]: healthz check failed Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.137525 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7p4xb" podUID="52a18549-2c1c-4f28-bfa1-6fb14e0690e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.153549 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:33 crc kubenswrapper[4787]: E1001 09:38:33.153892 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:33.653879219 +0000 UTC m=+145.769023376 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.157446 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-n47zx"] Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.186021 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-sgphp"] Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.204722 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cf76m"] Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.214199 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-9lvh8"] Oct 01 09:38:33 crc kubenswrapper[4787]: W1001 09:38:33.217547 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0f1d5b79_c7bd_4c71_aa2c_7e407c4b7adb.slice/crio-97ded80b61dac2ded0958013fd9bc5d2c1aa27a562f74a7908287a4a85f2a77f WatchSource:0}: Error finding container 97ded80b61dac2ded0958013fd9bc5d2c1aa27a562f74a7908287a4a85f2a77f: Status 404 returned error can't find the container with id 97ded80b61dac2ded0958013fd9bc5d2c1aa27a562f74a7908287a4a85f2a77f Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.247426 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-sgxp6"] Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.254397 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:33 crc kubenswrapper[4787]: E1001 09:38:33.254809 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:33.754785682 +0000 UTC m=+145.869929839 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.322903 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhp4" Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.322958 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhp4" Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.333875 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhp4" Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.357181 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:33 crc kubenswrapper[4787]: E1001 09:38:33.357671 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:33.857653415 +0000 UTC m=+145.972797572 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.469736 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:33 crc kubenswrapper[4787]: E1001 09:38:33.469799 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:33.969759845 +0000 UTC m=+146.084904002 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.485721 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:33 crc kubenswrapper[4787]: E1001 09:38:33.486205 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:33.986186986 +0000 UTC m=+146.101331143 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.587338 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:33 crc kubenswrapper[4787]: E1001 09:38:33.606312 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:34.10627118 +0000 UTC m=+146.221415337 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.606991 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:33 crc kubenswrapper[4787]: E1001 09:38:33.607758 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:34.107732958 +0000 UTC m=+146.222877115 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.708861 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:33 crc kubenswrapper[4787]: E1001 09:38:33.709442 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:34.209406351 +0000 UTC m=+146.324550508 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.709947 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:33 crc kubenswrapper[4787]: E1001 09:38:33.710418 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:34.210403365 +0000 UTC m=+146.325547522 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.810995 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:33 crc kubenswrapper[4787]: E1001 09:38:33.811189 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:34.311167095 +0000 UTC m=+146.426311252 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.811461 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:33 crc kubenswrapper[4787]: E1001 09:38:33.813462 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:34.312810388 +0000 UTC m=+146.427954545 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.844045 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qvdw4" event={"ID":"e1d30c03-586b-460e-be72-c1770024c8f6","Type":"ContainerStarted","Data":"83df95d3c01487ff4dd14fc48fdc14aa9632a32e5b7621645060e4c058138618"} Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.844123 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qvdw4" event={"ID":"e1d30c03-586b-460e-be72-c1770024c8f6","Type":"ContainerStarted","Data":"8e43aeb419d633375538c3474e0227ea57cfe7d31202c536c2ae661f211f87ec"} Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.852303 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-kw99x" event={"ID":"8de4a645-2945-4261-b9be-ee3336e818df","Type":"ContainerStarted","Data":"7c1e9231673196bcc39c382d8fed3c520fb5a11bde8abe81df02128bbaefbe48"} Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.852355 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-kw99x" event={"ID":"8de4a645-2945-4261-b9be-ee3336e818df","Type":"ContainerStarted","Data":"4bd1a34abf91ed08f32088d26f36a1ce8ad1e6a1b960b35b45aeda0de9a12b9c"} Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.868748 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qvdw4" podStartSLOduration=126.868728099 podStartE2EDuration="2m6.868728099s" podCreationTimestamp="2025-10-01 09:36:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:33.866295707 +0000 UTC m=+145.981439864" watchObservedRunningTime="2025-10-01 09:38:33.868728099 +0000 UTC m=+145.983872256" Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.894028 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-sgphp" event={"ID":"5c56c7aa-b794-474b-910b-11ecf8481009","Type":"ContainerStarted","Data":"b84196c0e1073bb0f109483ede44793113392c85b68f2d6b9c4edb4328aa2461"} Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.912379 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:33 crc kubenswrapper[4787]: E1001 09:38:33.912776 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:34.412750946 +0000 UTC m=+146.527895103 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.916700 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gjtg8" event={"ID":"82a8aff4-187a-464a-b375-5c20e576164d","Type":"ContainerStarted","Data":"f9d39144b6530de15e65187c54b86ba2339a380906657117842c9457dc4027d8"} Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.938463 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-kw99x" podStartSLOduration=125.938432403 podStartE2EDuration="2m5.938432403s" podCreationTimestamp="2025-10-01 09:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:33.904764621 +0000 UTC m=+146.019908778" watchObservedRunningTime="2025-10-01 09:38:33.938432403 +0000 UTC m=+146.053576560" Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.976631 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zssrz" event={"ID":"e6b988c0-936d-4b49-80a2-015ae390f262","Type":"ContainerStarted","Data":"52764d4206fe707c89308dece827cde4bcb8906f36ced74544221b6a785936c1"} Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.987749 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-hqx7l" event={"ID":"d438f633-0698-4c39-9c5a-71a57a6b3fbc","Type":"ContainerStarted","Data":"1654e397b93b78933937b48e94971951de221db873c561f5f7d6df85c1db1b03"} Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.987803 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-hqx7l" event={"ID":"d438f633-0698-4c39-9c5a-71a57a6b3fbc","Type":"ContainerStarted","Data":"2d03f6ff6547aa7b5c30ee15f292a076c1baf4099b741eaabf6e1f04b9615996"} Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.989048 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-hqx7l" Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.990973 4787 patch_prober.go:28] interesting pod/console-operator-58897d9998-hqx7l container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.33:8443/readyz\": dial tcp 10.217.0.33:8443: connect: connection refused" start-of-body= Oct 01 09:38:33 crc kubenswrapper[4787]: I1001 09:38:33.991028 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-hqx7l" podUID="d438f633-0698-4c39-9c5a-71a57a6b3fbc" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.33:8443/readyz\": dial tcp 10.217.0.33:8443: connect: connection refused" Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.004853 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9x4nl" event={"ID":"54dbcab2-1757-4f7d-b59c-128e257b5660","Type":"ContainerStarted","Data":"1ffb9ad4ce2a92157925adae370562c05f7f2bc4f79d61b66a3a6e4e2abad078"} Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.013094 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gjtg8" podStartSLOduration=126.013054174 podStartE2EDuration="2m6.013054174s" podCreationTimestamp="2025-10-01 09:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:33.941592824 +0000 UTC m=+146.056736991" watchObservedRunningTime="2025-10-01 09:38:34.013054174 +0000 UTC m=+146.128198331" Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.014052 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zssrz" podStartSLOduration=126.014047219 podStartE2EDuration="2m6.014047219s" podCreationTimestamp="2025-10-01 09:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:34.011219067 +0000 UTC m=+146.126363234" watchObservedRunningTime="2025-10-01 09:38:34.014047219 +0000 UTC m=+146.129191376" Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.017004 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:34 crc kubenswrapper[4787]: E1001 09:38:34.021168 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:34.521152381 +0000 UTC m=+146.636296538 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.037540 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pzdzn" event={"ID":"0ed372ed-3529-4e31-b124-d66f292089d0","Type":"ContainerStarted","Data":"6aec5417ecd21d55a156dcb5561a0a2508cfe9bf0bf0d9c54f3528d29e659cd1"} Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.039882 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-9lvh8" event={"ID":"9f2a2142-5430-4edb-8f20-b199d0425466","Type":"ContainerStarted","Data":"050324ad86e7e9cfd8ba35da2d06398fba92dd1e0680d0e16491ce8025fc1914"} Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.094919 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-blwqt" event={"ID":"13cafd2b-3d1e-4d8b-8751-a583af048423","Type":"ContainerStarted","Data":"0507b5f70db6f2c93461d297d873d82edf7f5472e9b1caeee863109c82b57cc4"} Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.094992 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-blwqt" event={"ID":"13cafd2b-3d1e-4d8b-8751-a583af048423","Type":"ContainerStarted","Data":"1cd4f49239bf5c7ee92b13eb53db46d09a555729ba820bb33d5d35521d7aaf41"} Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.100450 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-hqx7l" podStartSLOduration=126.10041994 podStartE2EDuration="2m6.10041994s" podCreationTimestamp="2025-10-01 09:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:34.04571586 +0000 UTC m=+146.160860017" watchObservedRunningTime="2025-10-01 09:38:34.10041994 +0000 UTC m=+146.215564097" Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.109664 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-whc4s" event={"ID":"556c69e8-1179-4de2-841d-3f096c3adb18","Type":"ContainerStarted","Data":"056c0b95dfe0e4b9c330db4cce41908e38ed33c46f9e3902d8f3132b857c8237"} Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.109719 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-whc4s" event={"ID":"556c69e8-1179-4de2-841d-3f096c3adb18","Type":"ContainerStarted","Data":"865605b6fb543b69e55d98063b717689fe8b9681898ab76270069c6d8eb863e2"} Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.111575 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-whc4s" Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.112964 4787 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-whc4s container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" start-of-body= Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.113000 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-whc4s" podUID="556c69e8-1179-4de2-841d-3f096c3adb18" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.118409 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:34 crc kubenswrapper[4787]: E1001 09:38:34.120036 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:34.620005762 +0000 UTC m=+146.735149919 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.123177 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-ncnkh" event={"ID":"151f124f-e071-4884-aeb4-217064807fa8","Type":"ContainerStarted","Data":"84b945f99d406a95a1530b8e8653913d646e4e967e1b0c3ec75cbd8a76900c99"} Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.123190 4787 patch_prober.go:28] interesting pod/router-default-5444994796-7p4xb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 09:38:34 crc kubenswrapper[4787]: [-]has-synced failed: reason withheld Oct 01 09:38:34 crc kubenswrapper[4787]: [+]process-running ok Oct 01 09:38:34 crc kubenswrapper[4787]: healthz check failed Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.123264 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7p4xb" podUID="52a18549-2c1c-4f28-bfa1-6fb14e0690e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.126615 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-gqvt9" event={"ID":"2ec1df2f-eadd-4385-97b4-664101fb9d51","Type":"ContainerStarted","Data":"a4c271522ac1c586ceb37ed14cec52855bc91429be6335c564b8a0318fb00173"} Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.126647 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-gqvt9" event={"ID":"2ec1df2f-eadd-4385-97b4-664101fb9d51","Type":"ContainerStarted","Data":"7fb44713d867d5f5b7e61272f9947cc5ddbb9c637b20236ca48d6ba01ac38bb3"} Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.130510 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-tn48z" event={"ID":"59fc7ef6-b52d-41e5-8c82-9ce4ed5d5053","Type":"ContainerStarted","Data":"aec78d794fc905a31bae971f5272e70585549574ff48de2af18d917ba6b83198"} Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.145185 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-whc4s" podStartSLOduration=126.145166796 podStartE2EDuration="2m6.145166796s" podCreationTimestamp="2025-10-01 09:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:34.142954119 +0000 UTC m=+146.258098276" watchObservedRunningTime="2025-10-01 09:38:34.145166796 +0000 UTC m=+146.260310953" Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.145852 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pzdzn" podStartSLOduration=127.145847903 podStartE2EDuration="2m7.145847903s" podCreationTimestamp="2025-10-01 09:36:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:34.097234919 +0000 UTC m=+146.212379076" watchObservedRunningTime="2025-10-01 09:38:34.145847903 +0000 UTC m=+146.260992060" Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.184920 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tzsx8" event={"ID":"fa4ba99a-cfa6-4e3b-bd6f-ce18ea3fefe8","Type":"ContainerStarted","Data":"c65b4dc1c629cdac91ad7228bf3e4f71b9fca0732ef7517d01b6a29138acb032"} Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.189339 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-gqvt9" podStartSLOduration=126.189317536 podStartE2EDuration="2m6.189317536s" podCreationTimestamp="2025-10-01 09:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:34.183813845 +0000 UTC m=+146.298958012" watchObservedRunningTime="2025-10-01 09:38:34.189317536 +0000 UTC m=+146.304461703" Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.215709 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n47zx" event={"ID":"0f1d5b79-c7bd-4c71-aa2c-7e407c4b7adb","Type":"ContainerStarted","Data":"97ded80b61dac2ded0958013fd9bc5d2c1aa27a562f74a7908287a4a85f2a77f"} Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.223410 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:34 crc kubenswrapper[4787]: E1001 09:38:34.225452 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:34.725438591 +0000 UTC m=+146.840582748 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.242633 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-86dvz" event={"ID":"e9338499-2f07-4087-b75f-bf85d8a2a794","Type":"ContainerStarted","Data":"50808a82f6cfab13a3b3041aa2d676d9396112da1d8ccc8623cccd14c61b1b74"} Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.259540 4787 generic.go:334] "Generic (PLEG): container finished" podID="98c8956e-c6c0-4e79-b7a5-64054256133a" containerID="fe3437cee35c8389c486e606da48b8dfa3f9adb499892d7235fbd4b85753d97c" exitCode=0 Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.259603 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-lld4x" event={"ID":"98c8956e-c6c0-4e79-b7a5-64054256133a","Type":"ContainerDied","Data":"fe3437cee35c8389c486e606da48b8dfa3f9adb499892d7235fbd4b85753d97c"} Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.259629 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-lld4x" event={"ID":"98c8956e-c6c0-4e79-b7a5-64054256133a","Type":"ContainerStarted","Data":"f2501c197ecda66a181bc2b847f4dd5d0ba33eded98c5c6fcc047940069d3ee8"} Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.293651 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-sgxp6" event={"ID":"a2b762fa-551a-40c6-acd8-45e6649949d6","Type":"ContainerStarted","Data":"dafade3fb608ba9eb8e0b864ca9cd0bdd9a0917e1a47cda7a2b2260554214aa2"} Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.294622 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tzsx8" podStartSLOduration=126.294602252 podStartE2EDuration="2m6.294602252s" podCreationTimestamp="2025-10-01 09:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:34.234686998 +0000 UTC m=+146.349831155" watchObservedRunningTime="2025-10-01 09:38:34.294602252 +0000 UTC m=+146.409746399" Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.294845 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-86dvz" podStartSLOduration=126.294840638 podStartE2EDuration="2m6.294840638s" podCreationTimestamp="2025-10-01 09:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:34.292835686 +0000 UTC m=+146.407979843" watchObservedRunningTime="2025-10-01 09:38:34.294840638 +0000 UTC m=+146.409984785" Oct 01 09:38:34 crc kubenswrapper[4787]: E1001 09:38:34.324997 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:34.824956338 +0000 UTC m=+146.940100495 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.324866 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.325773 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:34 crc kubenswrapper[4787]: E1001 09:38:34.328367 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:34.828344175 +0000 UTC m=+146.943488322 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.374535 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-sgxp6" podStartSLOduration=7.354049873 podStartE2EDuration="7.354049873s" podCreationTimestamp="2025-10-01 09:38:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:34.346370127 +0000 UTC m=+146.461514304" watchObservedRunningTime="2025-10-01 09:38:34.354049873 +0000 UTC m=+146.469194030" Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.404160 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-29mhg" event={"ID":"523e6f72-37ee-4c32-b3e0-98405775373f","Type":"ContainerStarted","Data":"b46fe9a7ad75c2f69de3672d9eecfd10e1218751ebe134704d96685b5a4f966d"} Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.429783 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:34 crc kubenswrapper[4787]: E1001 09:38:34.431537 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:34.931508285 +0000 UTC m=+147.046652442 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.443407 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-n6959" event={"ID":"75578ed4-3bb2-4ed7-bac1-d28780be4d67","Type":"ContainerStarted","Data":"b49a1a666e0cb22b80f402b2be1e57b2fd025b11920302f1eeba365fc5aaeac7"} Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.451166 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cf76m" event={"ID":"5d7d263b-bc04-4069-81a5-d5cd1151d805","Type":"ContainerStarted","Data":"ad1ffd81aa54a22f70acb1266afe5bb2e53d7b7561d798f32bdef1c2f7c3fd5c"} Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.523869 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-fsblz" event={"ID":"aca7d6cc-eca3-4ffb-84d9-254616d0c037","Type":"ContainerStarted","Data":"763cfb484c7152b28308fb696614ecbf30c6080a61ef0ec2a70efccd37eb225a"} Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.524257 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-fsblz" event={"ID":"aca7d6cc-eca3-4ffb-84d9-254616d0c037","Type":"ContainerStarted","Data":"c32e8d33d79150cf8ac08b055ebe31f092ed3ad74dba8d40eee5e0ea0e9c9934"} Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.542792 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:34 crc kubenswrapper[4787]: E1001 09:38:34.552386 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:35.05236349 +0000 UTC m=+147.167507647 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.563771 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-46vzw" event={"ID":"b0ea3010-e975-4d1e-9fa4-385489660daa","Type":"ContainerStarted","Data":"a3d5720c9a0c02c5c5aa43cea7fbec6eae843b882aa22ba35123ecf8ed08145e"} Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.578135 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r8xzr" event={"ID":"ee393336-510d-40f3-a382-06d7f839750f","Type":"ContainerStarted","Data":"750a84946dbe3fe0cdf83194942e0fe866775ed544c8a8d5b0a28862e778547b"} Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.582129 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r8xzr" Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.597589 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-29mhg" podStartSLOduration=126.597567367 podStartE2EDuration="2m6.597567367s" podCreationTimestamp="2025-10-01 09:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:34.432215224 +0000 UTC m=+146.547359381" watchObservedRunningTime="2025-10-01 09:38:34.597567367 +0000 UTC m=+146.712711524" Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.601861 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.605137 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rv6xx" Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.619045 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hlhp4" Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.624656 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-46vzw" podStartSLOduration=127.6246309 podStartE2EDuration="2m7.6246309s" podCreationTimestamp="2025-10-01 09:36:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:34.595607447 +0000 UTC m=+146.710751604" watchObservedRunningTime="2025-10-01 09:38:34.6246309 +0000 UTC m=+146.739775047" Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.652674 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r8xzr" podStartSLOduration=126.652655817 podStartE2EDuration="2m6.652655817s" podCreationTimestamp="2025-10-01 09:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:34.635900098 +0000 UTC m=+146.751044265" watchObservedRunningTime="2025-10-01 09:38:34.652655817 +0000 UTC m=+146.767799974" Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.666626 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:34 crc kubenswrapper[4787]: E1001 09:38:34.668227 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:35.168205965 +0000 UTC m=+147.283350122 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.774289 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:34 crc kubenswrapper[4787]: E1001 09:38:34.774953 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:35.274938048 +0000 UTC m=+147.390082205 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.877571 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:34 crc kubenswrapper[4787]: E1001 09:38:34.877977 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:35.377961815 +0000 UTC m=+147.493105972 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:34 crc kubenswrapper[4787]: I1001 09:38:34.978946 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:34 crc kubenswrapper[4787]: E1001 09:38:34.984134 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:35.484117283 +0000 UTC m=+147.599261440 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.089257 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:35 crc kubenswrapper[4787]: E1001 09:38:35.089991 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:35.589974153 +0000 UTC m=+147.705118310 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.120331 4787 patch_prober.go:28] interesting pod/router-default-5444994796-7p4xb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 09:38:35 crc kubenswrapper[4787]: [-]has-synced failed: reason withheld Oct 01 09:38:35 crc kubenswrapper[4787]: [+]process-running ok Oct 01 09:38:35 crc kubenswrapper[4787]: healthz check failed Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.120389 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7p4xb" podUID="52a18549-2c1c-4f28-bfa1-6fb14e0690e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.130051 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r8xzr" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.192420 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:35 crc kubenswrapper[4787]: E1001 09:38:35.192747 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:35.692735704 +0000 UTC m=+147.807879861 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.248114 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cqr6k"] Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.249762 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cqr6k" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.303369 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:35 crc kubenswrapper[4787]: E1001 09:38:35.303571 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:35.80354146 +0000 UTC m=+147.918685617 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.303896 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7fht\" (UniqueName: \"kubernetes.io/projected/c4fd105d-9feb-4898-92a1-9c7ae1613202-kube-api-access-s7fht\") pod \"certified-operators-cqr6k\" (UID: \"c4fd105d-9feb-4898-92a1-9c7ae1613202\") " pod="openshift-marketplace/certified-operators-cqr6k" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.304016 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4fd105d-9feb-4898-92a1-9c7ae1613202-catalog-content\") pod \"certified-operators-cqr6k\" (UID: \"c4fd105d-9feb-4898-92a1-9c7ae1613202\") " pod="openshift-marketplace/certified-operators-cqr6k" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.304121 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4fd105d-9feb-4898-92a1-9c7ae1613202-utilities\") pod \"certified-operators-cqr6k\" (UID: \"c4fd105d-9feb-4898-92a1-9c7ae1613202\") " pod="openshift-marketplace/certified-operators-cqr6k" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.304249 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:35 crc kubenswrapper[4787]: E1001 09:38:35.304619 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:35.804611208 +0000 UTC m=+147.919755365 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.306942 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.318608 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cqr6k"] Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.399222 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xq2kw"] Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.400194 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xq2kw" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.410066 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.410452 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.410689 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7fht\" (UniqueName: \"kubernetes.io/projected/c4fd105d-9feb-4898-92a1-9c7ae1613202-kube-api-access-s7fht\") pod \"certified-operators-cqr6k\" (UID: \"c4fd105d-9feb-4898-92a1-9c7ae1613202\") " pod="openshift-marketplace/certified-operators-cqr6k" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.410737 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4fd105d-9feb-4898-92a1-9c7ae1613202-catalog-content\") pod \"certified-operators-cqr6k\" (UID: \"c4fd105d-9feb-4898-92a1-9c7ae1613202\") " pod="openshift-marketplace/certified-operators-cqr6k" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.410759 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4fd105d-9feb-4898-92a1-9c7ae1613202-utilities\") pod \"certified-operators-cqr6k\" (UID: \"c4fd105d-9feb-4898-92a1-9c7ae1613202\") " pod="openshift-marketplace/certified-operators-cqr6k" Oct 01 09:38:35 crc kubenswrapper[4787]: E1001 09:38:35.411165 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:35.911127914 +0000 UTC m=+148.026272282 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.411428 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4fd105d-9feb-4898-92a1-9c7ae1613202-catalog-content\") pod \"certified-operators-cqr6k\" (UID: \"c4fd105d-9feb-4898-92a1-9c7ae1613202\") " pod="openshift-marketplace/certified-operators-cqr6k" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.414249 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4fd105d-9feb-4898-92a1-9c7ae1613202-utilities\") pod \"certified-operators-cqr6k\" (UID: \"c4fd105d-9feb-4898-92a1-9c7ae1613202\") " pod="openshift-marketplace/certified-operators-cqr6k" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.432821 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xq2kw"] Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.484740 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7fht\" (UniqueName: \"kubernetes.io/projected/c4fd105d-9feb-4898-92a1-9c7ae1613202-kube-api-access-s7fht\") pod \"certified-operators-cqr6k\" (UID: \"c4fd105d-9feb-4898-92a1-9c7ae1613202\") " pod="openshift-marketplace/certified-operators-cqr6k" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.512212 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/045fb03e-9b60-4729-b4cd-73db79bb6294-catalog-content\") pod \"community-operators-xq2kw\" (UID: \"045fb03e-9b60-4729-b4cd-73db79bb6294\") " pod="openshift-marketplace/community-operators-xq2kw" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.512585 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpzd6\" (UniqueName: \"kubernetes.io/projected/045fb03e-9b60-4729-b4cd-73db79bb6294-kube-api-access-gpzd6\") pod \"community-operators-xq2kw\" (UID: \"045fb03e-9b60-4729-b4cd-73db79bb6294\") " pod="openshift-marketplace/community-operators-xq2kw" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.512674 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/045fb03e-9b60-4729-b4cd-73db79bb6294-utilities\") pod \"community-operators-xq2kw\" (UID: \"045fb03e-9b60-4729-b4cd-73db79bb6294\") " pod="openshift-marketplace/community-operators-xq2kw" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.512808 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:35 crc kubenswrapper[4787]: E1001 09:38:35.513277 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:36.013261099 +0000 UTC m=+148.128405256 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.574522 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cqr6k" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.599099 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-n6959" event={"ID":"75578ed4-3bb2-4ed7-bac1-d28780be4d67","Type":"ContainerStarted","Data":"7fb2f45aa31a8c701110626dad0d939380e361b39779f8ff01f444917bf7c474"} Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.599474 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-n6959" event={"ID":"75578ed4-3bb2-4ed7-bac1-d28780be4d67","Type":"ContainerStarted","Data":"1217159881e3320d024152840369a5fc925114e6456313d2abc6cbbee0b75de7"} Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.609829 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-sgphp" event={"ID":"5c56c7aa-b794-474b-910b-11ecf8481009","Type":"ContainerStarted","Data":"58825ceb5c259da549d3f4ccbec95e57e78e77242d1cc642cd525509fe9aa6fe"} Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.620845 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:35 crc kubenswrapper[4787]: E1001 09:38:35.621166 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:36.121144101 +0000 UTC m=+148.236288258 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.621459 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.621815 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/045fb03e-9b60-4729-b4cd-73db79bb6294-catalog-content\") pod \"community-operators-xq2kw\" (UID: \"045fb03e-9b60-4729-b4cd-73db79bb6294\") " pod="openshift-marketplace/community-operators-xq2kw" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.621849 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpzd6\" (UniqueName: \"kubernetes.io/projected/045fb03e-9b60-4729-b4cd-73db79bb6294-kube-api-access-gpzd6\") pod \"community-operators-xq2kw\" (UID: \"045fb03e-9b60-4729-b4cd-73db79bb6294\") " pod="openshift-marketplace/community-operators-xq2kw" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.621876 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/045fb03e-9b60-4729-b4cd-73db79bb6294-utilities\") pod \"community-operators-xq2kw\" (UID: \"045fb03e-9b60-4729-b4cd-73db79bb6294\") " pod="openshift-marketplace/community-operators-xq2kw" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.622398 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/045fb03e-9b60-4729-b4cd-73db79bb6294-utilities\") pod \"community-operators-xq2kw\" (UID: \"045fb03e-9b60-4729-b4cd-73db79bb6294\") " pod="openshift-marketplace/community-operators-xq2kw" Oct 01 09:38:35 crc kubenswrapper[4787]: E1001 09:38:35.622689 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:36.12267968 +0000 UTC m=+148.237823837 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.622936 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/045fb03e-9b60-4729-b4cd-73db79bb6294-catalog-content\") pod \"community-operators-xq2kw\" (UID: \"045fb03e-9b60-4729-b4cd-73db79bb6294\") " pod="openshift-marketplace/community-operators-xq2kw" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.627264 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n47zx" event={"ID":"0f1d5b79-c7bd-4c71-aa2c-7e407c4b7adb","Type":"ContainerStarted","Data":"f24e8fa37b3c40cee9cc38f0bd856393617bba6e33302d40265077237cb97aec"} Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.627320 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n47zx" event={"ID":"0f1d5b79-c7bd-4c71-aa2c-7e407c4b7adb","Type":"ContainerStarted","Data":"4e11760e1cff3b02ccc204bcb764c711d554d178a328b6fc9942063a49b5325e"} Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.642157 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-n72ph"] Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.643071 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n72ph" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.644641 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-86dvz" event={"ID":"e9338499-2f07-4087-b75f-bf85d8a2a794","Type":"ContainerStarted","Data":"39aa4f53e303db5dc569c0834cbb0ff54b7f45aa83f1a3568a11c411e2a989fe"} Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.646044 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-29mhg" event={"ID":"523e6f72-37ee-4c32-b3e0-98405775373f","Type":"ContainerStarted","Data":"213f9cc316860ab525d5bddaa5113967d5660b8a0f8ecc73793a9e877e5a238f"} Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.671435 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpzd6\" (UniqueName: \"kubernetes.io/projected/045fb03e-9b60-4729-b4cd-73db79bb6294-kube-api-access-gpzd6\") pod \"community-operators-xq2kw\" (UID: \"045fb03e-9b60-4729-b4cd-73db79bb6294\") " pod="openshift-marketplace/community-operators-xq2kw" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.672594 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-n6959" podStartSLOduration=127.672576988 podStartE2EDuration="2m7.672576988s" podCreationTimestamp="2025-10-01 09:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:35.663917076 +0000 UTC m=+147.779061243" watchObservedRunningTime="2025-10-01 09:38:35.672576988 +0000 UTC m=+147.787721145" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.673351 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-n72ph"] Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.692304 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-tn48z" event={"ID":"59fc7ef6-b52d-41e5-8c82-9ce4ed5d5053","Type":"ContainerStarted","Data":"46f516a10e72a55a5a652a03c340f6a8fea94f435012d3793da67bef2c2ffe2f"} Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.694137 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-sgphp" podStartSLOduration=127.694109959 podStartE2EDuration="2m7.694109959s" podCreationTimestamp="2025-10-01 09:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:35.693396841 +0000 UTC m=+147.808540988" watchObservedRunningTime="2025-10-01 09:38:35.694109959 +0000 UTC m=+147.809254116" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.715444 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zssrz" event={"ID":"e6b988c0-936d-4b49-80a2-015ae390f262","Type":"ContainerStarted","Data":"a69af10529d69b75c508081efd711d33e735e49a7662b0cafa1850f9f123ea3e"} Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.724385 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.724858 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0637215a-7749-4b92-a60e-f24e2d1b1a29-utilities\") pod \"certified-operators-n72ph\" (UID: \"0637215a-7749-4b92-a60e-f24e2d1b1a29\") " pod="openshift-marketplace/certified-operators-n72ph" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.724902 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0637215a-7749-4b92-a60e-f24e2d1b1a29-catalog-content\") pod \"certified-operators-n72ph\" (UID: \"0637215a-7749-4b92-a60e-f24e2d1b1a29\") " pod="openshift-marketplace/certified-operators-n72ph" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.724948 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzx4v\" (UniqueName: \"kubernetes.io/projected/0637215a-7749-4b92-a60e-f24e2d1b1a29-kube-api-access-kzx4v\") pod \"certified-operators-n72ph\" (UID: \"0637215a-7749-4b92-a60e-f24e2d1b1a29\") " pod="openshift-marketplace/certified-operators-n72ph" Oct 01 09:38:35 crc kubenswrapper[4787]: E1001 09:38:35.726908 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:36.226877888 +0000 UTC m=+148.342022045 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.734629 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xq2kw" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.740374 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n47zx" podStartSLOduration=127.740353152 podStartE2EDuration="2m7.740353152s" podCreationTimestamp="2025-10-01 09:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:35.738338641 +0000 UTC m=+147.853482818" watchObservedRunningTime="2025-10-01 09:38:35.740353152 +0000 UTC m=+147.855497309" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.751771 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-9lvh8" event={"ID":"9f2a2142-5430-4edb-8f20-b199d0425466","Type":"ContainerStarted","Data":"5041463c210ebdd13fb6e43f8a250a69731bdc11fb43101ed8fe8638e781cc65"} Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.751844 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-9lvh8" event={"ID":"9f2a2142-5430-4edb-8f20-b199d0425466","Type":"ContainerStarted","Data":"660abf88ac381cf427ec0ac9e9aced7bde391bcb14fc8cd3b5020558ae9a76e1"} Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.752040 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-9lvh8" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.763321 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-sgxp6" event={"ID":"a2b762fa-551a-40c6-acd8-45e6649949d6","Type":"ContainerStarted","Data":"6b1f060ac4a07c547af3e36f23b88b26fbca12d73632ed928b3b5391c898d264"} Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.797593 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-lld4x" event={"ID":"98c8956e-c6c0-4e79-b7a5-64054256133a","Type":"ContainerStarted","Data":"b83d693ddec2cec02c39f19a478b9c00544be456b8cca9a01ea14e4b3fe53896"} Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.797973 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-lld4x" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.806132 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cf76m" event={"ID":"5d7d263b-bc04-4069-81a5-d5cd1151d805","Type":"ContainerStarted","Data":"412ca820d6e3487af9fa549927f8fcb2620256515d9da3df5360333631deaaf2"} Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.806894 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-cf76m" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.821535 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-fsblz" event={"ID":"aca7d6cc-eca3-4ffb-84d9-254616d0c037","Type":"ContainerStarted","Data":"4bea87631a1b1b1be350893b44ab2921dfd0f3e69bc0ab291ec2831811a5545b"} Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.834635 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jszln"] Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.839602 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0637215a-7749-4b92-a60e-f24e2d1b1a29-utilities\") pod \"certified-operators-n72ph\" (UID: \"0637215a-7749-4b92-a60e-f24e2d1b1a29\") " pod="openshift-marketplace/certified-operators-n72ph" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.839688 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0637215a-7749-4b92-a60e-f24e2d1b1a29-catalog-content\") pod \"certified-operators-n72ph\" (UID: \"0637215a-7749-4b92-a60e-f24e2d1b1a29\") " pod="openshift-marketplace/certified-operators-n72ph" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.839811 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzx4v\" (UniqueName: \"kubernetes.io/projected/0637215a-7749-4b92-a60e-f24e2d1b1a29-kube-api-access-kzx4v\") pod \"certified-operators-n72ph\" (UID: \"0637215a-7749-4b92-a60e-f24e2d1b1a29\") " pod="openshift-marketplace/certified-operators-n72ph" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.839839 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:35 crc kubenswrapper[4787]: E1001 09:38:35.842066 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:36.342046536 +0000 UTC m=+148.457190693 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.845786 4787 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-cf76m container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" start-of-body= Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.845878 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-cf76m" podUID="5d7d263b-bc04-4069-81a5-d5cd1151d805" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.847483 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0637215a-7749-4b92-a60e-f24e2d1b1a29-utilities\") pod \"certified-operators-n72ph\" (UID: \"0637215a-7749-4b92-a60e-f24e2d1b1a29\") " pod="openshift-marketplace/certified-operators-n72ph" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.855004 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0637215a-7749-4b92-a60e-f24e2d1b1a29-catalog-content\") pod \"certified-operators-n72ph\" (UID: \"0637215a-7749-4b92-a60e-f24e2d1b1a29\") " pod="openshift-marketplace/certified-operators-n72ph" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.856749 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jszln" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.865289 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jszln"] Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.904299 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-tn48z" podStartSLOduration=127.904270319 podStartE2EDuration="2m7.904270319s" podCreationTimestamp="2025-10-01 09:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:35.835493848 +0000 UTC m=+147.950638015" watchObservedRunningTime="2025-10-01 09:38:35.904270319 +0000 UTC m=+148.019414476" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.905949 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-9lvh8" podStartSLOduration=8.905942680999999 podStartE2EDuration="8.905942681s" podCreationTimestamp="2025-10-01 09:38:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:35.871647444 +0000 UTC m=+147.986791621" watchObservedRunningTime="2025-10-01 09:38:35.905942681 +0000 UTC m=+148.021086838" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.907992 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-blwqt" event={"ID":"13cafd2b-3d1e-4d8b-8751-a583af048423","Type":"ContainerStarted","Data":"8f57413212c685d5756492c6259c5831bfd3fe361af706ba72eee6004613e7a9"} Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.909105 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-blwqt" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.910492 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzx4v\" (UniqueName: \"kubernetes.io/projected/0637215a-7749-4b92-a60e-f24e2d1b1a29-kube-api-access-kzx4v\") pod \"certified-operators-n72ph\" (UID: \"0637215a-7749-4b92-a60e-f24e2d1b1a29\") " pod="openshift-marketplace/certified-operators-n72ph" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.942656 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.943057 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34c3f769-241a-4df9-9f8d-48308997f85c-utilities\") pod \"community-operators-jszln\" (UID: \"34c3f769-241a-4df9-9f8d-48308997f85c\") " pod="openshift-marketplace/community-operators-jszln" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.943112 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34c3f769-241a-4df9-9f8d-48308997f85c-catalog-content\") pod \"community-operators-jszln\" (UID: \"34c3f769-241a-4df9-9f8d-48308997f85c\") " pod="openshift-marketplace/community-operators-jszln" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.943169 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k587l\" (UniqueName: \"kubernetes.io/projected/34c3f769-241a-4df9-9f8d-48308997f85c-kube-api-access-k587l\") pod \"community-operators-jszln\" (UID: \"34c3f769-241a-4df9-9f8d-48308997f85c\") " pod="openshift-marketplace/community-operators-jszln" Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.943662 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-lld4x" podStartSLOduration=127.943639716 podStartE2EDuration="2m7.943639716s" podCreationTimestamp="2025-10-01 09:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:35.9425865 +0000 UTC m=+148.057730657" watchObservedRunningTime="2025-10-01 09:38:35.943639716 +0000 UTC m=+148.058783873" Oct 01 09:38:35 crc kubenswrapper[4787]: E1001 09:38:35.946282 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:36.446252663 +0000 UTC m=+148.561396810 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.946483 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-ncnkh" event={"ID":"151f124f-e071-4884-aeb4-217064807fa8","Type":"ContainerStarted","Data":"ce56ce1e0f670a3876b10cd866dd433e274f606f58ea56bb77d3dbf193598e25"} Oct 01 09:38:35 crc kubenswrapper[4787]: I1001 09:38:35.991319 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9x4nl" event={"ID":"54dbcab2-1757-4f7d-b59c-128e257b5660","Type":"ContainerStarted","Data":"de7f0c3f7237278866b86d7cbff16ca46e353187126fd192ed906d547d2b470f"} Oct 01 09:38:36 crc kubenswrapper[4787]: I1001 09:38:36.009295 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n72ph" Oct 01 09:38:36 crc kubenswrapper[4787]: I1001 09:38:36.021807 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-whc4s" Oct 01 09:38:36 crc kubenswrapper[4787]: I1001 09:38:36.045943 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-fsblz" podStartSLOduration=128.045913764 podStartE2EDuration="2m8.045913764s" podCreationTimestamp="2025-10-01 09:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:36.014356517 +0000 UTC m=+148.129500684" watchObservedRunningTime="2025-10-01 09:38:36.045913764 +0000 UTC m=+148.161057921" Oct 01 09:38:36 crc kubenswrapper[4787]: I1001 09:38:36.048339 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k587l\" (UniqueName: \"kubernetes.io/projected/34c3f769-241a-4df9-9f8d-48308997f85c-kube-api-access-k587l\") pod \"community-operators-jszln\" (UID: \"34c3f769-241a-4df9-9f8d-48308997f85c\") " pod="openshift-marketplace/community-operators-jszln" Oct 01 09:38:36 crc kubenswrapper[4787]: I1001 09:38:36.049023 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:36 crc kubenswrapper[4787]: I1001 09:38:36.049121 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34c3f769-241a-4df9-9f8d-48308997f85c-utilities\") pod \"community-operators-jszln\" (UID: \"34c3f769-241a-4df9-9f8d-48308997f85c\") " pod="openshift-marketplace/community-operators-jszln" Oct 01 09:38:36 crc kubenswrapper[4787]: I1001 09:38:36.049253 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34c3f769-241a-4df9-9f8d-48308997f85c-catalog-content\") pod \"community-operators-jszln\" (UID: \"34c3f769-241a-4df9-9f8d-48308997f85c\") " pod="openshift-marketplace/community-operators-jszln" Oct 01 09:38:36 crc kubenswrapper[4787]: I1001 09:38:36.049498 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-cf76m" podStartSLOduration=128.049491026 podStartE2EDuration="2m8.049491026s" podCreationTimestamp="2025-10-01 09:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:36.046389387 +0000 UTC m=+148.161533564" watchObservedRunningTime="2025-10-01 09:38:36.049491026 +0000 UTC m=+148.164635183" Oct 01 09:38:36 crc kubenswrapper[4787]: I1001 09:38:36.049797 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34c3f769-241a-4df9-9f8d-48308997f85c-catalog-content\") pod \"community-operators-jszln\" (UID: \"34c3f769-241a-4df9-9f8d-48308997f85c\") " pod="openshift-marketplace/community-operators-jszln" Oct 01 09:38:36 crc kubenswrapper[4787]: E1001 09:38:36.062562 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:36.56254301 +0000 UTC m=+148.677687157 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:36 crc kubenswrapper[4787]: I1001 09:38:36.062847 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34c3f769-241a-4df9-9f8d-48308997f85c-utilities\") pod \"community-operators-jszln\" (UID: \"34c3f769-241a-4df9-9f8d-48308997f85c\") " pod="openshift-marketplace/community-operators-jszln" Oct 01 09:38:36 crc kubenswrapper[4787]: I1001 09:38:36.102071 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9x4nl" podStartSLOduration=128.102044932 podStartE2EDuration="2m8.102044932s" podCreationTimestamp="2025-10-01 09:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:36.101669332 +0000 UTC m=+148.216813499" watchObservedRunningTime="2025-10-01 09:38:36.102044932 +0000 UTC m=+148.217189089" Oct 01 09:38:36 crc kubenswrapper[4787]: I1001 09:38:36.152941 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:36 crc kubenswrapper[4787]: E1001 09:38:36.154029 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:36.653992832 +0000 UTC m=+148.769136989 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:36 crc kubenswrapper[4787]: I1001 09:38:36.176354 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k587l\" (UniqueName: \"kubernetes.io/projected/34c3f769-241a-4df9-9f8d-48308997f85c-kube-api-access-k587l\") pod \"community-operators-jszln\" (UID: \"34c3f769-241a-4df9-9f8d-48308997f85c\") " pod="openshift-marketplace/community-operators-jszln" Oct 01 09:38:36 crc kubenswrapper[4787]: I1001 09:38:36.184211 4787 patch_prober.go:28] interesting pod/router-default-5444994796-7p4xb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 09:38:36 crc kubenswrapper[4787]: [-]has-synced failed: reason withheld Oct 01 09:38:36 crc kubenswrapper[4787]: [+]process-running ok Oct 01 09:38:36 crc kubenswrapper[4787]: healthz check failed Oct 01 09:38:36 crc kubenswrapper[4787]: I1001 09:38:36.184282 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7p4xb" podUID="52a18549-2c1c-4f28-bfa1-6fb14e0690e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 09:38:36 crc kubenswrapper[4787]: I1001 09:38:36.256483 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:36 crc kubenswrapper[4787]: E1001 09:38:36.256875 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:36.756858645 +0000 UTC m=+148.872002802 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:36 crc kubenswrapper[4787]: I1001 09:38:36.282596 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jszln" Oct 01 09:38:36 crc kubenswrapper[4787]: I1001 09:38:36.359032 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:36 crc kubenswrapper[4787]: E1001 09:38:36.359466 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:36.859441561 +0000 UTC m=+148.974585718 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:36 crc kubenswrapper[4787]: I1001 09:38:36.461725 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:36 crc kubenswrapper[4787]: I1001 09:38:36.461801 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:38:36 crc kubenswrapper[4787]: I1001 09:38:36.461886 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:38:36 crc kubenswrapper[4787]: I1001 09:38:36.461930 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:38:36 crc kubenswrapper[4787]: I1001 09:38:36.461959 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:38:36 crc kubenswrapper[4787]: I1001 09:38:36.468648 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:38:36 crc kubenswrapper[4787]: E1001 09:38:36.469112 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:36.969067718 +0000 UTC m=+149.084211875 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:36 crc kubenswrapper[4787]: I1001 09:38:36.473770 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:38:36 crc kubenswrapper[4787]: I1001 09:38:36.476160 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:38:36 crc kubenswrapper[4787]: I1001 09:38:36.485106 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:38:36 crc kubenswrapper[4787]: I1001 09:38:36.485487 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-blwqt" podStartSLOduration=128.485474478 podStartE2EDuration="2m8.485474478s" podCreationTimestamp="2025-10-01 09:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:36.231747542 +0000 UTC m=+148.346891689" watchObservedRunningTime="2025-10-01 09:38:36.485474478 +0000 UTC m=+148.600618635" Oct 01 09:38:36 crc kubenswrapper[4787]: I1001 09:38:36.487147 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cqr6k"] Oct 01 09:38:36 crc kubenswrapper[4787]: I1001 09:38:36.543650 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 01 09:38:36 crc kubenswrapper[4787]: I1001 09:38:36.555236 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 01 09:38:36 crc kubenswrapper[4787]: I1001 09:38:36.563326 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:36 crc kubenswrapper[4787]: E1001 09:38:36.563846 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:37.063819023 +0000 UTC m=+149.178963180 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:36 crc kubenswrapper[4787]: I1001 09:38:36.568140 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xq2kw"] Oct 01 09:38:36 crc kubenswrapper[4787]: I1001 09:38:36.576159 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-hqx7l" Oct 01 09:38:36 crc kubenswrapper[4787]: I1001 09:38:36.638510 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:38:36 crc kubenswrapper[4787]: I1001 09:38:36.665838 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:36 crc kubenswrapper[4787]: E1001 09:38:36.666366 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:37.166352149 +0000 UTC m=+149.281496306 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:36 crc kubenswrapper[4787]: I1001 09:38:36.772280 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:36 crc kubenswrapper[4787]: E1001 09:38:36.772714 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:37.27269732 +0000 UTC m=+149.387841477 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:36 crc kubenswrapper[4787]: I1001 09:38:36.874145 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:36 crc kubenswrapper[4787]: E1001 09:38:36.875185 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:37.375161494 +0000 UTC m=+149.490305641 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:36 crc kubenswrapper[4787]: I1001 09:38:36.925026 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jszln"] Oct 01 09:38:36 crc kubenswrapper[4787]: I1001 09:38:36.976454 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:36 crc kubenswrapper[4787]: E1001 09:38:36.976937 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:37.476920268 +0000 UTC m=+149.592064425 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.007275 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-n72ph"] Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.026514 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xq2kw" event={"ID":"045fb03e-9b60-4729-b4cd-73db79bb6294","Type":"ContainerStarted","Data":"e62b65509b8f464082ef0e2f8df708bb9c15c57f13257e30d8bc208e746b154a"} Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.039466 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cqr6k" event={"ID":"c4fd105d-9feb-4898-92a1-9c7ae1613202","Type":"ContainerStarted","Data":"77064cc1d6b633d8c53eeaf5d1957a67a410973e670c0c8c0115072b30a81249"} Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.065371 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jszln" event={"ID":"34c3f769-241a-4df9-9f8d-48308997f85c","Type":"ContainerStarted","Data":"47ee8912dc6b6ba20c6471d3c149f8c238814cdb20bcefa278fb9f500565273d"} Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.078006 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:37 crc kubenswrapper[4787]: E1001 09:38:37.078325 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:37.578314795 +0000 UTC m=+149.693458952 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.083335 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-ncnkh" event={"ID":"151f124f-e071-4884-aeb4-217064807fa8","Type":"ContainerStarted","Data":"41b0fbcc685293e74221365ee687f48fabc908d167734eed8715b8dd53a6d035"} Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.095104 4787 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-cf76m container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" start-of-body= Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.095148 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-cf76m" podUID="5d7d263b-bc04-4069-81a5-d5cd1151d805" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.117924 4787 patch_prober.go:28] interesting pod/router-default-5444994796-7p4xb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 09:38:37 crc kubenswrapper[4787]: [-]has-synced failed: reason withheld Oct 01 09:38:37 crc kubenswrapper[4787]: [+]process-running ok Oct 01 09:38:37 crc kubenswrapper[4787]: healthz check failed Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.117976 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7p4xb" podUID="52a18549-2c1c-4f28-bfa1-6fb14e0690e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.179070 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:37 crc kubenswrapper[4787]: E1001 09:38:37.180965 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:37.680948522 +0000 UTC m=+149.796092679 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.213339 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-z2fsb"] Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.219154 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z2fsb" Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.237940 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.280908 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/315c2b9e-c2cc-4192-8a74-fe35860ff463-utilities\") pod \"redhat-marketplace-z2fsb\" (UID: \"315c2b9e-c2cc-4192-8a74-fe35860ff463\") " pod="openshift-marketplace/redhat-marketplace-z2fsb" Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.280968 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/315c2b9e-c2cc-4192-8a74-fe35860ff463-catalog-content\") pod \"redhat-marketplace-z2fsb\" (UID: \"315c2b9e-c2cc-4192-8a74-fe35860ff463\") " pod="openshift-marketplace/redhat-marketplace-z2fsb" Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.281005 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m96f7\" (UniqueName: \"kubernetes.io/projected/315c2b9e-c2cc-4192-8a74-fe35860ff463-kube-api-access-m96f7\") pod \"redhat-marketplace-z2fsb\" (UID: \"315c2b9e-c2cc-4192-8a74-fe35860ff463\") " pod="openshift-marketplace/redhat-marketplace-z2fsb" Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.281044 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:37 crc kubenswrapper[4787]: E1001 09:38:37.281393 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:37.781381333 +0000 UTC m=+149.896525490 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.308437 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z2fsb"] Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.383324 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:37 crc kubenswrapper[4787]: E1001 09:38:37.383541 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:37.883503237 +0000 UTC m=+149.998647384 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.383599 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/315c2b9e-c2cc-4192-8a74-fe35860ff463-catalog-content\") pod \"redhat-marketplace-z2fsb\" (UID: \"315c2b9e-c2cc-4192-8a74-fe35860ff463\") " pod="openshift-marketplace/redhat-marketplace-z2fsb" Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.383698 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m96f7\" (UniqueName: \"kubernetes.io/projected/315c2b9e-c2cc-4192-8a74-fe35860ff463-kube-api-access-m96f7\") pod \"redhat-marketplace-z2fsb\" (UID: \"315c2b9e-c2cc-4192-8a74-fe35860ff463\") " pod="openshift-marketplace/redhat-marketplace-z2fsb" Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.383801 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.384002 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/315c2b9e-c2cc-4192-8a74-fe35860ff463-utilities\") pod \"redhat-marketplace-z2fsb\" (UID: \"315c2b9e-c2cc-4192-8a74-fe35860ff463\") " pod="openshift-marketplace/redhat-marketplace-z2fsb" Oct 01 09:38:37 crc kubenswrapper[4787]: E1001 09:38:37.386503 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:37.886480673 +0000 UTC m=+150.001624820 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.386731 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/315c2b9e-c2cc-4192-8a74-fe35860ff463-utilities\") pod \"redhat-marketplace-z2fsb\" (UID: \"315c2b9e-c2cc-4192-8a74-fe35860ff463\") " pod="openshift-marketplace/redhat-marketplace-z2fsb" Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.386914 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/315c2b9e-c2cc-4192-8a74-fe35860ff463-catalog-content\") pod \"redhat-marketplace-z2fsb\" (UID: \"315c2b9e-c2cc-4192-8a74-fe35860ff463\") " pod="openshift-marketplace/redhat-marketplace-z2fsb" Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.442051 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m96f7\" (UniqueName: \"kubernetes.io/projected/315c2b9e-c2cc-4192-8a74-fe35860ff463-kube-api-access-m96f7\") pod \"redhat-marketplace-z2fsb\" (UID: \"315c2b9e-c2cc-4192-8a74-fe35860ff463\") " pod="openshift-marketplace/redhat-marketplace-z2fsb" Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.495566 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:37 crc kubenswrapper[4787]: E1001 09:38:37.496513 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:37.996488989 +0000 UTC m=+150.111633136 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.498923 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-lld4x" Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.552173 4787 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.588096 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tzk9m"] Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.589654 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tzk9m" Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.599901 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:37 crc kubenswrapper[4787]: E1001 09:38:37.600273 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-01 09:38:38.100256427 +0000 UTC m=+150.215400584 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cpqnn" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.610035 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tzk9m"] Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.701231 4787 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-01T09:38:37.552206076Z","Handler":null,"Name":""} Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.702300 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.702785 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa6719df-be6a-4209-9f40-70fbba0465a1-catalog-content\") pod \"redhat-marketplace-tzk9m\" (UID: \"aa6719df-be6a-4209-9f40-70fbba0465a1\") " pod="openshift-marketplace/redhat-marketplace-tzk9m" Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.702874 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvxmq\" (UniqueName: \"kubernetes.io/projected/aa6719df-be6a-4209-9f40-70fbba0465a1-kube-api-access-jvxmq\") pod \"redhat-marketplace-tzk9m\" (UID: \"aa6719df-be6a-4209-9f40-70fbba0465a1\") " pod="openshift-marketplace/redhat-marketplace-tzk9m" Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.702941 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa6719df-be6a-4209-9f40-70fbba0465a1-utilities\") pod \"redhat-marketplace-tzk9m\" (UID: \"aa6719df-be6a-4209-9f40-70fbba0465a1\") " pod="openshift-marketplace/redhat-marketplace-tzk9m" Oct 01 09:38:37 crc kubenswrapper[4787]: E1001 09:38:37.703131 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-01 09:38:38.203078388 +0000 UTC m=+150.318222545 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.707728 4787 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.707771 4787 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.732524 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z2fsb" Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.805440 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa6719df-be6a-4209-9f40-70fbba0465a1-utilities\") pod \"redhat-marketplace-tzk9m\" (UID: \"aa6719df-be6a-4209-9f40-70fbba0465a1\") " pod="openshift-marketplace/redhat-marketplace-tzk9m" Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.805975 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa6719df-be6a-4209-9f40-70fbba0465a1-catalog-content\") pod \"redhat-marketplace-tzk9m\" (UID: \"aa6719df-be6a-4209-9f40-70fbba0465a1\") " pod="openshift-marketplace/redhat-marketplace-tzk9m" Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.806036 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.806107 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvxmq\" (UniqueName: \"kubernetes.io/projected/aa6719df-be6a-4209-9f40-70fbba0465a1-kube-api-access-jvxmq\") pod \"redhat-marketplace-tzk9m\" (UID: \"aa6719df-be6a-4209-9f40-70fbba0465a1\") " pod="openshift-marketplace/redhat-marketplace-tzk9m" Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.807343 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa6719df-be6a-4209-9f40-70fbba0465a1-catalog-content\") pod \"redhat-marketplace-tzk9m\" (UID: \"aa6719df-be6a-4209-9f40-70fbba0465a1\") " pod="openshift-marketplace/redhat-marketplace-tzk9m" Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.807680 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa6719df-be6a-4209-9f40-70fbba0465a1-utilities\") pod \"redhat-marketplace-tzk9m\" (UID: \"aa6719df-be6a-4209-9f40-70fbba0465a1\") " pod="openshift-marketplace/redhat-marketplace-tzk9m" Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.815846 4787 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.815912 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.831508 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvxmq\" (UniqueName: \"kubernetes.io/projected/aa6719df-be6a-4209-9f40-70fbba0465a1-kube-api-access-jvxmq\") pod \"redhat-marketplace-tzk9m\" (UID: \"aa6719df-be6a-4209-9f40-70fbba0465a1\") " pod="openshift-marketplace/redhat-marketplace-tzk9m" Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.856524 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cpqnn\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.919939 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.921227 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tzk9m" Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.957967 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 01 09:38:37 crc kubenswrapper[4787]: I1001 09:38:37.978982 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z2fsb"] Oct 01 09:38:38 crc kubenswrapper[4787]: W1001 09:38:38.000871 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod315c2b9e_c2cc_4192_8a74_fe35860ff463.slice/crio-382db25a103803d1ab0d9d3dea0c7e40a10df32e543b3365854a351972ffac73 WatchSource:0}: Error finding container 382db25a103803d1ab0d9d3dea0c7e40a10df32e543b3365854a351972ffac73: Status 404 returned error can't find the container with id 382db25a103803d1ab0d9d3dea0c7e40a10df32e543b3365854a351972ffac73 Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.068571 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.092962 4787 generic.go:334] "Generic (PLEG): container finished" podID="c4fd105d-9feb-4898-92a1-9c7ae1613202" containerID="ba462dbf29972e8af033234af97d7fca424f7e6f2f9296e4a1fe27e84e570c98" exitCode=0 Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.093020 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cqr6k" event={"ID":"c4fd105d-9feb-4898-92a1-9c7ae1613202","Type":"ContainerDied","Data":"ba462dbf29972e8af033234af97d7fca424f7e6f2f9296e4a1fe27e84e570c98"} Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.099255 4787 generic.go:334] "Generic (PLEG): container finished" podID="0637215a-7749-4b92-a60e-f24e2d1b1a29" containerID="992bf39e914a8b5f27b7448fd98046bf25404b4039770bca150fbfdf6f29d667" exitCode=0 Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.099385 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n72ph" event={"ID":"0637215a-7749-4b92-a60e-f24e2d1b1a29","Type":"ContainerDied","Data":"992bf39e914a8b5f27b7448fd98046bf25404b4039770bca150fbfdf6f29d667"} Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.099730 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n72ph" event={"ID":"0637215a-7749-4b92-a60e-f24e2d1b1a29","Type":"ContainerStarted","Data":"4ff46a5fc2376146771db7fb30bfeb45d340acf1aaf46139a7c1124624387a96"} Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.104381 4787 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.108164 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"fe7b9f2e8a45f83bd2e73e19e26a1a4119c63a7aea97b5dd0207eba1b2cd9828"} Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.108235 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"96a7a94a5e9464dab1a8f775a8ea6535d2a3118835c78a28b59265e579f7893d"} Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.120372 4787 patch_prober.go:28] interesting pod/router-default-5444994796-7p4xb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 09:38:38 crc kubenswrapper[4787]: [-]has-synced failed: reason withheld Oct 01 09:38:38 crc kubenswrapper[4787]: [+]process-running ok Oct 01 09:38:38 crc kubenswrapper[4787]: healthz check failed Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.120464 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7p4xb" podUID="52a18549-2c1c-4f28-bfa1-6fb14e0690e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.127253 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"af29da9fcf23252f06492959b9dda1af2706bcf5e14b65bf34fcf3a4e6acf9a0"} Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.127294 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"79c9871a2401e0b58b31abf35e7f569d516760ff29b327046fd46d4743fbd919"} Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.127571 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.132929 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"58e0117856b425c2ea37e47d25c0a978829841705cf0c927dd92a70ae9421444"} Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.132982 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"5a243b48ebc63bc58b17d62ae2017a9e9a2b5acdc290b4aebd272f20a03ed7c7"} Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.149204 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-ncnkh" event={"ID":"151f124f-e071-4884-aeb4-217064807fa8","Type":"ContainerStarted","Data":"8e06b1c78073d25e90c20650d9838fc0ec139a7611296fc9168f6e45d91ebe10"} Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.152128 4787 generic.go:334] "Generic (PLEG): container finished" podID="045fb03e-9b60-4729-b4cd-73db79bb6294" containerID="e440081149f9e943d08a9704521ed7cd0ad44c46e0a65bed4170f50a62fd7e78" exitCode=0 Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.152251 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xq2kw" event={"ID":"045fb03e-9b60-4729-b4cd-73db79bb6294","Type":"ContainerDied","Data":"e440081149f9e943d08a9704521ed7cd0ad44c46e0a65bed4170f50a62fd7e78"} Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.161292 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z2fsb" event={"ID":"315c2b9e-c2cc-4192-8a74-fe35860ff463","Type":"ContainerStarted","Data":"382db25a103803d1ab0d9d3dea0c7e40a10df32e543b3365854a351972ffac73"} Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.166988 4787 generic.go:334] "Generic (PLEG): container finished" podID="34c3f769-241a-4df9-9f8d-48308997f85c" containerID="a0d13b015eb0406ac50c47a4c50c7ebbeb73445272303b0c75282e8cc39357d4" exitCode=0 Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.167729 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jszln" event={"ID":"34c3f769-241a-4df9-9f8d-48308997f85c","Type":"ContainerDied","Data":"a0d13b015eb0406ac50c47a4c50c7ebbeb73445272303b0c75282e8cc39357d4"} Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.206361 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tzk9m"] Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.544828 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.545997 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cpqnn"] Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.583716 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-t6rc9"] Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.584867 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t6rc9" Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.587744 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.591544 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t6rc9"] Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.742667 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4914dc93-bedf-4429-a399-9f465e2c0ab9-utilities\") pod \"redhat-operators-t6rc9\" (UID: \"4914dc93-bedf-4429-a399-9f465e2c0ab9\") " pod="openshift-marketplace/redhat-operators-t6rc9" Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.742715 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4914dc93-bedf-4429-a399-9f465e2c0ab9-catalog-content\") pod \"redhat-operators-t6rc9\" (UID: \"4914dc93-bedf-4429-a399-9f465e2c0ab9\") " pod="openshift-marketplace/redhat-operators-t6rc9" Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.742793 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4dxt\" (UniqueName: \"kubernetes.io/projected/4914dc93-bedf-4429-a399-9f465e2c0ab9-kube-api-access-b4dxt\") pod \"redhat-operators-t6rc9\" (UID: \"4914dc93-bedf-4429-a399-9f465e2c0ab9\") " pod="openshift-marketplace/redhat-operators-t6rc9" Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.843983 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4914dc93-bedf-4429-a399-9f465e2c0ab9-catalog-content\") pod \"redhat-operators-t6rc9\" (UID: \"4914dc93-bedf-4429-a399-9f465e2c0ab9\") " pod="openshift-marketplace/redhat-operators-t6rc9" Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.844137 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4dxt\" (UniqueName: \"kubernetes.io/projected/4914dc93-bedf-4429-a399-9f465e2c0ab9-kube-api-access-b4dxt\") pod \"redhat-operators-t6rc9\" (UID: \"4914dc93-bedf-4429-a399-9f465e2c0ab9\") " pod="openshift-marketplace/redhat-operators-t6rc9" Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.844227 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4914dc93-bedf-4429-a399-9f465e2c0ab9-utilities\") pod \"redhat-operators-t6rc9\" (UID: \"4914dc93-bedf-4429-a399-9f465e2c0ab9\") " pod="openshift-marketplace/redhat-operators-t6rc9" Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.844727 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4914dc93-bedf-4429-a399-9f465e2c0ab9-catalog-content\") pod \"redhat-operators-t6rc9\" (UID: \"4914dc93-bedf-4429-a399-9f465e2c0ab9\") " pod="openshift-marketplace/redhat-operators-t6rc9" Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.844999 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4914dc93-bedf-4429-a399-9f465e2c0ab9-utilities\") pod \"redhat-operators-t6rc9\" (UID: \"4914dc93-bedf-4429-a399-9f465e2c0ab9\") " pod="openshift-marketplace/redhat-operators-t6rc9" Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.873342 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4dxt\" (UniqueName: \"kubernetes.io/projected/4914dc93-bedf-4429-a399-9f465e2c0ab9-kube-api-access-b4dxt\") pod \"redhat-operators-t6rc9\" (UID: \"4914dc93-bedf-4429-a399-9f465e2c0ab9\") " pod="openshift-marketplace/redhat-operators-t6rc9" Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.940713 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t6rc9" Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.985271 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-p5759"] Oct 01 09:38:38 crc kubenswrapper[4787]: I1001 09:38:38.986872 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p5759" Oct 01 09:38:39 crc kubenswrapper[4787]: I1001 09:38:38.993325 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p5759"] Oct 01 09:38:39 crc kubenswrapper[4787]: I1001 09:38:39.110502 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-7p4xb" Oct 01 09:38:39 crc kubenswrapper[4787]: I1001 09:38:39.115853 4787 patch_prober.go:28] interesting pod/router-default-5444994796-7p4xb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 09:38:39 crc kubenswrapper[4787]: [-]has-synced failed: reason withheld Oct 01 09:38:39 crc kubenswrapper[4787]: [+]process-running ok Oct 01 09:38:39 crc kubenswrapper[4787]: healthz check failed Oct 01 09:38:39 crc kubenswrapper[4787]: I1001 09:38:39.116189 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7p4xb" podUID="52a18549-2c1c-4f28-bfa1-6fb14e0690e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 09:38:39 crc kubenswrapper[4787]: I1001 09:38:39.148967 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8hlk\" (UniqueName: \"kubernetes.io/projected/33bdb919-2c9f-4133-be8b-368dea5b099a-kube-api-access-g8hlk\") pod \"redhat-operators-p5759\" (UID: \"33bdb919-2c9f-4133-be8b-368dea5b099a\") " pod="openshift-marketplace/redhat-operators-p5759" Oct 01 09:38:39 crc kubenswrapper[4787]: I1001 09:38:39.149047 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33bdb919-2c9f-4133-be8b-368dea5b099a-catalog-content\") pod \"redhat-operators-p5759\" (UID: \"33bdb919-2c9f-4133-be8b-368dea5b099a\") " pod="openshift-marketplace/redhat-operators-p5759" Oct 01 09:38:39 crc kubenswrapper[4787]: I1001 09:38:39.149105 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33bdb919-2c9f-4133-be8b-368dea5b099a-utilities\") pod \"redhat-operators-p5759\" (UID: \"33bdb919-2c9f-4133-be8b-368dea5b099a\") " pod="openshift-marketplace/redhat-operators-p5759" Oct 01 09:38:39 crc kubenswrapper[4787]: I1001 09:38:39.178757 4787 generic.go:334] "Generic (PLEG): container finished" podID="aa6719df-be6a-4209-9f40-70fbba0465a1" containerID="3b327333988d481b0c9efcac604b3b5b61c97afd9d63c6548eda57f11da0f523" exitCode=0 Oct 01 09:38:39 crc kubenswrapper[4787]: I1001 09:38:39.178833 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tzk9m" event={"ID":"aa6719df-be6a-4209-9f40-70fbba0465a1","Type":"ContainerDied","Data":"3b327333988d481b0c9efcac604b3b5b61c97afd9d63c6548eda57f11da0f523"} Oct 01 09:38:39 crc kubenswrapper[4787]: I1001 09:38:39.178869 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tzk9m" event={"ID":"aa6719df-be6a-4209-9f40-70fbba0465a1","Type":"ContainerStarted","Data":"b0bf6ed1803f786ade43847557e4c9f42883fee4f25e3f39c57abe637df4e7b1"} Oct 01 09:38:39 crc kubenswrapper[4787]: I1001 09:38:39.184329 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" event={"ID":"8ffe49e7-9382-4711-985c-8d6f76c42abc","Type":"ContainerStarted","Data":"24e6211e9d0aba2dcda0020a7110833544f47433f0c35f6d9ce9aad377589a16"} Oct 01 09:38:39 crc kubenswrapper[4787]: I1001 09:38:39.184380 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" event={"ID":"8ffe49e7-9382-4711-985c-8d6f76c42abc","Type":"ContainerStarted","Data":"924be146bfef052c7afcc388ecd46b6c52f34b9ef3c62a14b479dc715baca166"} Oct 01 09:38:39 crc kubenswrapper[4787]: I1001 09:38:39.184688 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:39 crc kubenswrapper[4787]: I1001 09:38:39.201250 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-ncnkh" event={"ID":"151f124f-e071-4884-aeb4-217064807fa8","Type":"ContainerStarted","Data":"0b5110f6c174c624a9b97b131ba22a26adc1c30a088c9aa4258fb3883a3c0c52"} Oct 01 09:38:39 crc kubenswrapper[4787]: I1001 09:38:39.208734 4787 generic.go:334] "Generic (PLEG): container finished" podID="315c2b9e-c2cc-4192-8a74-fe35860ff463" containerID="107f12d0b4ce8bec9f0bbc58687db3f5e3662af34817e98d5b27c328da014fa2" exitCode=0 Oct 01 09:38:39 crc kubenswrapper[4787]: I1001 09:38:39.209048 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z2fsb" event={"ID":"315c2b9e-c2cc-4192-8a74-fe35860ff463","Type":"ContainerDied","Data":"107f12d0b4ce8bec9f0bbc58687db3f5e3662af34817e98d5b27c328da014fa2"} Oct 01 09:38:39 crc kubenswrapper[4787]: I1001 09:38:39.263872 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33bdb919-2c9f-4133-be8b-368dea5b099a-utilities\") pod \"redhat-operators-p5759\" (UID: \"33bdb919-2c9f-4133-be8b-368dea5b099a\") " pod="openshift-marketplace/redhat-operators-p5759" Oct 01 09:38:39 crc kubenswrapper[4787]: I1001 09:38:39.264033 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8hlk\" (UniqueName: \"kubernetes.io/projected/33bdb919-2c9f-4133-be8b-368dea5b099a-kube-api-access-g8hlk\") pod \"redhat-operators-p5759\" (UID: \"33bdb919-2c9f-4133-be8b-368dea5b099a\") " pod="openshift-marketplace/redhat-operators-p5759" Oct 01 09:38:39 crc kubenswrapper[4787]: I1001 09:38:39.264102 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33bdb919-2c9f-4133-be8b-368dea5b099a-catalog-content\") pod \"redhat-operators-p5759\" (UID: \"33bdb919-2c9f-4133-be8b-368dea5b099a\") " pod="openshift-marketplace/redhat-operators-p5759" Oct 01 09:38:39 crc kubenswrapper[4787]: I1001 09:38:39.265061 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33bdb919-2c9f-4133-be8b-368dea5b099a-catalog-content\") pod \"redhat-operators-p5759\" (UID: \"33bdb919-2c9f-4133-be8b-368dea5b099a\") " pod="openshift-marketplace/redhat-operators-p5759" Oct 01 09:38:39 crc kubenswrapper[4787]: I1001 09:38:39.266731 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33bdb919-2c9f-4133-be8b-368dea5b099a-utilities\") pod \"redhat-operators-p5759\" (UID: \"33bdb919-2c9f-4133-be8b-368dea5b099a\") " pod="openshift-marketplace/redhat-operators-p5759" Oct 01 09:38:39 crc kubenswrapper[4787]: I1001 09:38:39.286936 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" podStartSLOduration=131.286908014 podStartE2EDuration="2m11.286908014s" podCreationTimestamp="2025-10-01 09:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:39.249071236 +0000 UTC m=+151.364215393" watchObservedRunningTime="2025-10-01 09:38:39.286908014 +0000 UTC m=+151.402052171" Oct 01 09:38:39 crc kubenswrapper[4787]: I1001 09:38:39.288687 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t6rc9"] Oct 01 09:38:39 crc kubenswrapper[4787]: I1001 09:38:39.303038 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-ncnkh" podStartSLOduration=12.303011497 podStartE2EDuration="12.303011497s" podCreationTimestamp="2025-10-01 09:38:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:39.296364227 +0000 UTC m=+151.411508414" watchObservedRunningTime="2025-10-01 09:38:39.303011497 +0000 UTC m=+151.418155654" Oct 01 09:38:39 crc kubenswrapper[4787]: I1001 09:38:39.324178 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8hlk\" (UniqueName: \"kubernetes.io/projected/33bdb919-2c9f-4133-be8b-368dea5b099a-kube-api-access-g8hlk\") pod \"redhat-operators-p5759\" (UID: \"33bdb919-2c9f-4133-be8b-368dea5b099a\") " pod="openshift-marketplace/redhat-operators-p5759" Oct 01 09:38:39 crc kubenswrapper[4787]: I1001 09:38:39.326210 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p5759" Oct 01 09:38:39 crc kubenswrapper[4787]: W1001 09:38:39.332442 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4914dc93_bedf_4429_a399_9f465e2c0ab9.slice/crio-72fef4d89d69491f191f474d85600c9ef288790c967779d310570c0bed16819d WatchSource:0}: Error finding container 72fef4d89d69491f191f474d85600c9ef288790c967779d310570c0bed16819d: Status 404 returned error can't find the container with id 72fef4d89d69491f191f474d85600c9ef288790c967779d310570c0bed16819d Oct 01 09:38:39 crc kubenswrapper[4787]: I1001 09:38:39.450589 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-46vzw" Oct 01 09:38:39 crc kubenswrapper[4787]: I1001 09:38:39.451938 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-46vzw" Oct 01 09:38:39 crc kubenswrapper[4787]: I1001 09:38:39.468401 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-46vzw" Oct 01 09:38:39 crc kubenswrapper[4787]: I1001 09:38:39.649202 4787 patch_prober.go:28] interesting pod/downloads-7954f5f757-tc862 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Oct 01 09:38:39 crc kubenswrapper[4787]: I1001 09:38:39.649863 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-tc862" podUID="b160cc9e-253c-4aba-951a-21fd20ff52f4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Oct 01 09:38:39 crc kubenswrapper[4787]: I1001 09:38:39.650696 4787 patch_prober.go:28] interesting pod/downloads-7954f5f757-tc862 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Oct 01 09:38:39 crc kubenswrapper[4787]: I1001 09:38:39.650758 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-tc862" podUID="b160cc9e-253c-4aba-951a-21fd20ff52f4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Oct 01 09:38:39 crc kubenswrapper[4787]: I1001 09:38:39.753035 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p5759"] Oct 01 09:38:39 crc kubenswrapper[4787]: I1001 09:38:39.915134 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-gqvt9" Oct 01 09:38:39 crc kubenswrapper[4787]: I1001 09:38:39.915520 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-gqvt9" Oct 01 09:38:39 crc kubenswrapper[4787]: I1001 09:38:39.918361 4787 patch_prober.go:28] interesting pod/console-f9d7485db-gqvt9 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.24:8443/health\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Oct 01 09:38:39 crc kubenswrapper[4787]: I1001 09:38:39.918423 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-gqvt9" podUID="2ec1df2f-eadd-4385-97b4-664101fb9d51" containerName="console" probeResult="failure" output="Get \"https://10.217.0.24:8443/health\": dial tcp 10.217.0.24:8443: connect: connection refused" Oct 01 09:38:39 crc kubenswrapper[4787]: I1001 09:38:39.962665 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-cf76m" Oct 01 09:38:40 crc kubenswrapper[4787]: I1001 09:38:40.115746 4787 patch_prober.go:28] interesting pod/router-default-5444994796-7p4xb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 09:38:40 crc kubenswrapper[4787]: [-]has-synced failed: reason withheld Oct 01 09:38:40 crc kubenswrapper[4787]: [+]process-running ok Oct 01 09:38:40 crc kubenswrapper[4787]: healthz check failed Oct 01 09:38:40 crc kubenswrapper[4787]: I1001 09:38:40.115845 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7p4xb" podUID="52a18549-2c1c-4f28-bfa1-6fb14e0690e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 09:38:40 crc kubenswrapper[4787]: I1001 09:38:40.229241 4787 generic.go:334] "Generic (PLEG): container finished" podID="4914dc93-bedf-4429-a399-9f465e2c0ab9" containerID="5671c7654e169dd5e940a2d4607a2cc33bc39e383de818bb73a97c4bc3fece1b" exitCode=0 Oct 01 09:38:40 crc kubenswrapper[4787]: I1001 09:38:40.229340 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t6rc9" event={"ID":"4914dc93-bedf-4429-a399-9f465e2c0ab9","Type":"ContainerDied","Data":"5671c7654e169dd5e940a2d4607a2cc33bc39e383de818bb73a97c4bc3fece1b"} Oct 01 09:38:40 crc kubenswrapper[4787]: I1001 09:38:40.229404 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t6rc9" event={"ID":"4914dc93-bedf-4429-a399-9f465e2c0ab9","Type":"ContainerStarted","Data":"72fef4d89d69491f191f474d85600c9ef288790c967779d310570c0bed16819d"} Oct 01 09:38:40 crc kubenswrapper[4787]: I1001 09:38:40.233475 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5759" event={"ID":"33bdb919-2c9f-4133-be8b-368dea5b099a","Type":"ContainerStarted","Data":"9737f3591d2c84e57b5e6fd3b421a8b9ac9aaa395227e2da1442a4ff1c592a6c"} Oct 01 09:38:40 crc kubenswrapper[4787]: I1001 09:38:40.233507 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5759" event={"ID":"33bdb919-2c9f-4133-be8b-368dea5b099a","Type":"ContainerStarted","Data":"d96d47b1ab4af429e19124794a21315419157ddbd78f5109f7e42c6c2ce4d4d9"} Oct 01 09:38:40 crc kubenswrapper[4787]: I1001 09:38:40.238426 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-46vzw" Oct 01 09:38:40 crc kubenswrapper[4787]: I1001 09:38:40.666989 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 01 09:38:40 crc kubenswrapper[4787]: I1001 09:38:40.667683 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 09:38:40 crc kubenswrapper[4787]: I1001 09:38:40.671109 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 01 09:38:40 crc kubenswrapper[4787]: I1001 09:38:40.671514 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 01 09:38:40 crc kubenswrapper[4787]: I1001 09:38:40.685687 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 01 09:38:40 crc kubenswrapper[4787]: I1001 09:38:40.764064 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:38:40 crc kubenswrapper[4787]: I1001 09:38:40.828164 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/aa71891b-473a-4cfd-a80a-802a90f14ce4-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"aa71891b-473a-4cfd-a80a-802a90f14ce4\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 09:38:40 crc kubenswrapper[4787]: I1001 09:38:40.828257 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/aa71891b-473a-4cfd-a80a-802a90f14ce4-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"aa71891b-473a-4cfd-a80a-802a90f14ce4\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 09:38:40 crc kubenswrapper[4787]: I1001 09:38:40.931055 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/aa71891b-473a-4cfd-a80a-802a90f14ce4-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"aa71891b-473a-4cfd-a80a-802a90f14ce4\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 09:38:40 crc kubenswrapper[4787]: I1001 09:38:40.931170 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/aa71891b-473a-4cfd-a80a-802a90f14ce4-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"aa71891b-473a-4cfd-a80a-802a90f14ce4\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 09:38:40 crc kubenswrapper[4787]: I1001 09:38:40.931246 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/aa71891b-473a-4cfd-a80a-802a90f14ce4-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"aa71891b-473a-4cfd-a80a-802a90f14ce4\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 09:38:40 crc kubenswrapper[4787]: I1001 09:38:40.962962 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/aa71891b-473a-4cfd-a80a-802a90f14ce4-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"aa71891b-473a-4cfd-a80a-802a90f14ce4\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 09:38:40 crc kubenswrapper[4787]: I1001 09:38:40.998393 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 09:38:41 crc kubenswrapper[4787]: I1001 09:38:41.118160 4787 patch_prober.go:28] interesting pod/router-default-5444994796-7p4xb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 09:38:41 crc kubenswrapper[4787]: [-]has-synced failed: reason withheld Oct 01 09:38:41 crc kubenswrapper[4787]: [+]process-running ok Oct 01 09:38:41 crc kubenswrapper[4787]: healthz check failed Oct 01 09:38:41 crc kubenswrapper[4787]: I1001 09:38:41.118232 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7p4xb" podUID="52a18549-2c1c-4f28-bfa1-6fb14e0690e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 09:38:41 crc kubenswrapper[4787]: I1001 09:38:41.251683 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:38:41 crc kubenswrapper[4787]: I1001 09:38:41.251809 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:38:41 crc kubenswrapper[4787]: I1001 09:38:41.277685 4787 generic.go:334] "Generic (PLEG): container finished" podID="33bdb919-2c9f-4133-be8b-368dea5b099a" containerID="9737f3591d2c84e57b5e6fd3b421a8b9ac9aaa395227e2da1442a4ff1c592a6c" exitCode=0 Oct 01 09:38:41 crc kubenswrapper[4787]: I1001 09:38:41.277768 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5759" event={"ID":"33bdb919-2c9f-4133-be8b-368dea5b099a","Type":"ContainerDied","Data":"9737f3591d2c84e57b5e6fd3b421a8b9ac9aaa395227e2da1442a4ff1c592a6c"} Oct 01 09:38:41 crc kubenswrapper[4787]: I1001 09:38:41.764555 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 01 09:38:42 crc kubenswrapper[4787]: I1001 09:38:42.116330 4787 patch_prober.go:28] interesting pod/router-default-5444994796-7p4xb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 09:38:42 crc kubenswrapper[4787]: [-]has-synced failed: reason withheld Oct 01 09:38:42 crc kubenswrapper[4787]: [+]process-running ok Oct 01 09:38:42 crc kubenswrapper[4787]: healthz check failed Oct 01 09:38:42 crc kubenswrapper[4787]: I1001 09:38:42.116400 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7p4xb" podUID="52a18549-2c1c-4f28-bfa1-6fb14e0690e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 09:38:42 crc kubenswrapper[4787]: I1001 09:38:42.289945 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"aa71891b-473a-4cfd-a80a-802a90f14ce4","Type":"ContainerStarted","Data":"d4d89e0e7456e929dd504dc2cf7de3e68090165bea8f9b553c79783619bd3989"} Oct 01 09:38:43 crc kubenswrapper[4787]: I1001 09:38:43.113990 4787 patch_prober.go:28] interesting pod/router-default-5444994796-7p4xb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 09:38:43 crc kubenswrapper[4787]: [-]has-synced failed: reason withheld Oct 01 09:38:43 crc kubenswrapper[4787]: [+]process-running ok Oct 01 09:38:43 crc kubenswrapper[4787]: healthz check failed Oct 01 09:38:43 crc kubenswrapper[4787]: I1001 09:38:43.114308 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7p4xb" podUID="52a18549-2c1c-4f28-bfa1-6fb14e0690e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 09:38:43 crc kubenswrapper[4787]: I1001 09:38:43.306608 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"aa71891b-473a-4cfd-a80a-802a90f14ce4","Type":"ContainerStarted","Data":"ebd3091033a18782b36cf8a53748c729ef62dff5cf2292030bcd461d89f56d98"} Oct 01 09:38:43 crc kubenswrapper[4787]: I1001 09:38:43.317001 4787 generic.go:334] "Generic (PLEG): container finished" podID="8de4a645-2945-4261-b9be-ee3336e818df" containerID="7c1e9231673196bcc39c382d8fed3c520fb5a11bde8abe81df02128bbaefbe48" exitCode=0 Oct 01 09:38:43 crc kubenswrapper[4787]: I1001 09:38:43.317047 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-kw99x" event={"ID":"8de4a645-2945-4261-b9be-ee3336e818df","Type":"ContainerDied","Data":"7c1e9231673196bcc39c382d8fed3c520fb5a11bde8abe81df02128bbaefbe48"} Oct 01 09:38:43 crc kubenswrapper[4787]: I1001 09:38:43.326860 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=3.326841968 podStartE2EDuration="3.326841968s" podCreationTimestamp="2025-10-01 09:38:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:38:43.321240674 +0000 UTC m=+155.436384831" watchObservedRunningTime="2025-10-01 09:38:43.326841968 +0000 UTC m=+155.441986125" Oct 01 09:38:43 crc kubenswrapper[4787]: I1001 09:38:43.405589 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 01 09:38:43 crc kubenswrapper[4787]: I1001 09:38:43.406499 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 09:38:43 crc kubenswrapper[4787]: I1001 09:38:43.408749 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 01 09:38:43 crc kubenswrapper[4787]: I1001 09:38:43.408925 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 01 09:38:43 crc kubenswrapper[4787]: I1001 09:38:43.424321 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 01 09:38:43 crc kubenswrapper[4787]: I1001 09:38:43.586722 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9ec992e3-443f-4601-9faf-ca18c832a8db-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9ec992e3-443f-4601-9faf-ca18c832a8db\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 09:38:43 crc kubenswrapper[4787]: I1001 09:38:43.586787 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9ec992e3-443f-4601-9faf-ca18c832a8db-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9ec992e3-443f-4601-9faf-ca18c832a8db\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 09:38:43 crc kubenswrapper[4787]: I1001 09:38:43.688790 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9ec992e3-443f-4601-9faf-ca18c832a8db-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9ec992e3-443f-4601-9faf-ca18c832a8db\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 09:38:43 crc kubenswrapper[4787]: I1001 09:38:43.688891 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9ec992e3-443f-4601-9faf-ca18c832a8db-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9ec992e3-443f-4601-9faf-ca18c832a8db\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 09:38:43 crc kubenswrapper[4787]: I1001 09:38:43.688977 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9ec992e3-443f-4601-9faf-ca18c832a8db-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9ec992e3-443f-4601-9faf-ca18c832a8db\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 09:38:43 crc kubenswrapper[4787]: I1001 09:38:43.838251 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9ec992e3-443f-4601-9faf-ca18c832a8db-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9ec992e3-443f-4601-9faf-ca18c832a8db\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 09:38:44 crc kubenswrapper[4787]: I1001 09:38:44.078592 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 09:38:44 crc kubenswrapper[4787]: I1001 09:38:44.113775 4787 patch_prober.go:28] interesting pod/router-default-5444994796-7p4xb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 09:38:44 crc kubenswrapper[4787]: [-]has-synced failed: reason withheld Oct 01 09:38:44 crc kubenswrapper[4787]: [+]process-running ok Oct 01 09:38:44 crc kubenswrapper[4787]: healthz check failed Oct 01 09:38:44 crc kubenswrapper[4787]: I1001 09:38:44.113837 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7p4xb" podUID="52a18549-2c1c-4f28-bfa1-6fb14e0690e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 09:38:44 crc kubenswrapper[4787]: I1001 09:38:44.346003 4787 generic.go:334] "Generic (PLEG): container finished" podID="aa71891b-473a-4cfd-a80a-802a90f14ce4" containerID="ebd3091033a18782b36cf8a53748c729ef62dff5cf2292030bcd461d89f56d98" exitCode=0 Oct 01 09:38:44 crc kubenswrapper[4787]: I1001 09:38:44.346240 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"aa71891b-473a-4cfd-a80a-802a90f14ce4","Type":"ContainerDied","Data":"ebd3091033a18782b36cf8a53748c729ef62dff5cf2292030bcd461d89f56d98"} Oct 01 09:38:44 crc kubenswrapper[4787]: I1001 09:38:44.435236 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 01 09:38:44 crc kubenswrapper[4787]: W1001 09:38:44.477752 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod9ec992e3_443f_4601_9faf_ca18c832a8db.slice/crio-b0e60a24fbbeab813f76e771c903db9d4b46a67e4fb43b8ebb74a76d6fb94660 WatchSource:0}: Error finding container b0e60a24fbbeab813f76e771c903db9d4b46a67e4fb43b8ebb74a76d6fb94660: Status 404 returned error can't find the container with id b0e60a24fbbeab813f76e771c903db9d4b46a67e4fb43b8ebb74a76d6fb94660 Oct 01 09:38:44 crc kubenswrapper[4787]: I1001 09:38:44.722168 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-kw99x" Oct 01 09:38:44 crc kubenswrapper[4787]: I1001 09:38:44.909002 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8de4a645-2945-4261-b9be-ee3336e818df-config-volume\") pod \"8de4a645-2945-4261-b9be-ee3336e818df\" (UID: \"8de4a645-2945-4261-b9be-ee3336e818df\") " Oct 01 09:38:44 crc kubenswrapper[4787]: I1001 09:38:44.909108 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6dm6j\" (UniqueName: \"kubernetes.io/projected/8de4a645-2945-4261-b9be-ee3336e818df-kube-api-access-6dm6j\") pod \"8de4a645-2945-4261-b9be-ee3336e818df\" (UID: \"8de4a645-2945-4261-b9be-ee3336e818df\") " Oct 01 09:38:44 crc kubenswrapper[4787]: I1001 09:38:44.909176 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8de4a645-2945-4261-b9be-ee3336e818df-secret-volume\") pod \"8de4a645-2945-4261-b9be-ee3336e818df\" (UID: \"8de4a645-2945-4261-b9be-ee3336e818df\") " Oct 01 09:38:44 crc kubenswrapper[4787]: I1001 09:38:44.910686 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8de4a645-2945-4261-b9be-ee3336e818df-config-volume" (OuterVolumeSpecName: "config-volume") pod "8de4a645-2945-4261-b9be-ee3336e818df" (UID: "8de4a645-2945-4261-b9be-ee3336e818df"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:38:44 crc kubenswrapper[4787]: I1001 09:38:44.930833 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8de4a645-2945-4261-b9be-ee3336e818df-kube-api-access-6dm6j" (OuterVolumeSpecName: "kube-api-access-6dm6j") pod "8de4a645-2945-4261-b9be-ee3336e818df" (UID: "8de4a645-2945-4261-b9be-ee3336e818df"). InnerVolumeSpecName "kube-api-access-6dm6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:38:44 crc kubenswrapper[4787]: I1001 09:38:44.933164 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8de4a645-2945-4261-b9be-ee3336e818df-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8de4a645-2945-4261-b9be-ee3336e818df" (UID: "8de4a645-2945-4261-b9be-ee3336e818df"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:38:45 crc kubenswrapper[4787]: I1001 09:38:45.013247 4787 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8de4a645-2945-4261-b9be-ee3336e818df-config-volume\") on node \"crc\" DevicePath \"\"" Oct 01 09:38:45 crc kubenswrapper[4787]: I1001 09:38:45.013289 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6dm6j\" (UniqueName: \"kubernetes.io/projected/8de4a645-2945-4261-b9be-ee3336e818df-kube-api-access-6dm6j\") on node \"crc\" DevicePath \"\"" Oct 01 09:38:45 crc kubenswrapper[4787]: I1001 09:38:45.013299 4787 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8de4a645-2945-4261-b9be-ee3336e818df-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 01 09:38:45 crc kubenswrapper[4787]: I1001 09:38:45.114558 4787 patch_prober.go:28] interesting pod/router-default-5444994796-7p4xb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 09:38:45 crc kubenswrapper[4787]: [-]has-synced failed: reason withheld Oct 01 09:38:45 crc kubenswrapper[4787]: [+]process-running ok Oct 01 09:38:45 crc kubenswrapper[4787]: healthz check failed Oct 01 09:38:45 crc kubenswrapper[4787]: I1001 09:38:45.114651 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7p4xb" podUID="52a18549-2c1c-4f28-bfa1-6fb14e0690e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 09:38:45 crc kubenswrapper[4787]: I1001 09:38:45.287231 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-9lvh8" Oct 01 09:38:45 crc kubenswrapper[4787]: I1001 09:38:45.379028 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9ec992e3-443f-4601-9faf-ca18c832a8db","Type":"ContainerStarted","Data":"b0e60a24fbbeab813f76e771c903db9d4b46a67e4fb43b8ebb74a76d6fb94660"} Oct 01 09:38:45 crc kubenswrapper[4787]: I1001 09:38:45.389462 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-kw99x" Oct 01 09:38:45 crc kubenswrapper[4787]: I1001 09:38:45.389493 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321850-kw99x" event={"ID":"8de4a645-2945-4261-b9be-ee3336e818df","Type":"ContainerDied","Data":"4bd1a34abf91ed08f32088d26f36a1ce8ad1e6a1b960b35b45aeda0de9a12b9c"} Oct 01 09:38:45 crc kubenswrapper[4787]: I1001 09:38:45.389547 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4bd1a34abf91ed08f32088d26f36a1ce8ad1e6a1b960b35b45aeda0de9a12b9c" Oct 01 09:38:45 crc kubenswrapper[4787]: I1001 09:38:45.756438 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 09:38:45 crc kubenswrapper[4787]: I1001 09:38:45.936647 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/aa71891b-473a-4cfd-a80a-802a90f14ce4-kube-api-access\") pod \"aa71891b-473a-4cfd-a80a-802a90f14ce4\" (UID: \"aa71891b-473a-4cfd-a80a-802a90f14ce4\") " Oct 01 09:38:45 crc kubenswrapper[4787]: I1001 09:38:45.936840 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/aa71891b-473a-4cfd-a80a-802a90f14ce4-kubelet-dir\") pod \"aa71891b-473a-4cfd-a80a-802a90f14ce4\" (UID: \"aa71891b-473a-4cfd-a80a-802a90f14ce4\") " Oct 01 09:38:45 crc kubenswrapper[4787]: I1001 09:38:45.936989 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aa71891b-473a-4cfd-a80a-802a90f14ce4-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "aa71891b-473a-4cfd-a80a-802a90f14ce4" (UID: "aa71891b-473a-4cfd-a80a-802a90f14ce4"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:38:45 crc kubenswrapper[4787]: I1001 09:38:45.937123 4787 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/aa71891b-473a-4cfd-a80a-802a90f14ce4-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 01 09:38:45 crc kubenswrapper[4787]: I1001 09:38:45.944274 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa71891b-473a-4cfd-a80a-802a90f14ce4-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "aa71891b-473a-4cfd-a80a-802a90f14ce4" (UID: "aa71891b-473a-4cfd-a80a-802a90f14ce4"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:38:46 crc kubenswrapper[4787]: I1001 09:38:46.041120 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/aa71891b-473a-4cfd-a80a-802a90f14ce4-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 01 09:38:46 crc kubenswrapper[4787]: I1001 09:38:46.113315 4787 patch_prober.go:28] interesting pod/router-default-5444994796-7p4xb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 09:38:46 crc kubenswrapper[4787]: [-]has-synced failed: reason withheld Oct 01 09:38:46 crc kubenswrapper[4787]: [+]process-running ok Oct 01 09:38:46 crc kubenswrapper[4787]: healthz check failed Oct 01 09:38:46 crc kubenswrapper[4787]: I1001 09:38:46.113409 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7p4xb" podUID="52a18549-2c1c-4f28-bfa1-6fb14e0690e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 09:38:46 crc kubenswrapper[4787]: I1001 09:38:46.434113 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"aa71891b-473a-4cfd-a80a-802a90f14ce4","Type":"ContainerDied","Data":"d4d89e0e7456e929dd504dc2cf7de3e68090165bea8f9b553c79783619bd3989"} Oct 01 09:38:46 crc kubenswrapper[4787]: I1001 09:38:46.434580 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d4d89e0e7456e929dd504dc2cf7de3e68090165bea8f9b553c79783619bd3989" Oct 01 09:38:46 crc kubenswrapper[4787]: I1001 09:38:46.434148 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 01 09:38:46 crc kubenswrapper[4787]: I1001 09:38:46.450490 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9ec992e3-443f-4601-9faf-ca18c832a8db","Type":"ContainerDied","Data":"a8aabbce0ae22a25604191df3b419dac432988fcb4e87474ed814e1b7f41260a"} Oct 01 09:38:46 crc kubenswrapper[4787]: I1001 09:38:46.450449 4787 generic.go:334] "Generic (PLEG): container finished" podID="9ec992e3-443f-4601-9faf-ca18c832a8db" containerID="a8aabbce0ae22a25604191df3b419dac432988fcb4e87474ed814e1b7f41260a" exitCode=0 Oct 01 09:38:47 crc kubenswrapper[4787]: I1001 09:38:47.112846 4787 patch_prober.go:28] interesting pod/router-default-5444994796-7p4xb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 09:38:47 crc kubenswrapper[4787]: [-]has-synced failed: reason withheld Oct 01 09:38:47 crc kubenswrapper[4787]: [+]process-running ok Oct 01 09:38:47 crc kubenswrapper[4787]: healthz check failed Oct 01 09:38:47 crc kubenswrapper[4787]: I1001 09:38:47.112957 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7p4xb" podUID="52a18549-2c1c-4f28-bfa1-6fb14e0690e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 09:38:48 crc kubenswrapper[4787]: I1001 09:38:48.113653 4787 patch_prober.go:28] interesting pod/router-default-5444994796-7p4xb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 09:38:48 crc kubenswrapper[4787]: [-]has-synced failed: reason withheld Oct 01 09:38:48 crc kubenswrapper[4787]: [+]process-running ok Oct 01 09:38:48 crc kubenswrapper[4787]: healthz check failed Oct 01 09:38:48 crc kubenswrapper[4787]: I1001 09:38:48.114383 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7p4xb" podUID="52a18549-2c1c-4f28-bfa1-6fb14e0690e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 09:38:49 crc kubenswrapper[4787]: I1001 09:38:49.113052 4787 patch_prober.go:28] interesting pod/router-default-5444994796-7p4xb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 09:38:49 crc kubenswrapper[4787]: [-]has-synced failed: reason withheld Oct 01 09:38:49 crc kubenswrapper[4787]: [+]process-running ok Oct 01 09:38:49 crc kubenswrapper[4787]: healthz check failed Oct 01 09:38:49 crc kubenswrapper[4787]: I1001 09:38:49.113343 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7p4xb" podUID="52a18549-2c1c-4f28-bfa1-6fb14e0690e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 09:38:49 crc kubenswrapper[4787]: I1001 09:38:49.657913 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-tc862" Oct 01 09:38:49 crc kubenswrapper[4787]: I1001 09:38:49.914839 4787 patch_prober.go:28] interesting pod/console-f9d7485db-gqvt9 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.24:8443/health\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Oct 01 09:38:49 crc kubenswrapper[4787]: I1001 09:38:49.914910 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-gqvt9" podUID="2ec1df2f-eadd-4385-97b4-664101fb9d51" containerName="console" probeResult="failure" output="Get \"https://10.217.0.24:8443/health\": dial tcp 10.217.0.24:8443: connect: connection refused" Oct 01 09:38:50 crc kubenswrapper[4787]: I1001 09:38:50.114070 4787 patch_prober.go:28] interesting pod/router-default-5444994796-7p4xb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 01 09:38:50 crc kubenswrapper[4787]: [-]has-synced failed: reason withheld Oct 01 09:38:50 crc kubenswrapper[4787]: [+]process-running ok Oct 01 09:38:50 crc kubenswrapper[4787]: healthz check failed Oct 01 09:38:50 crc kubenswrapper[4787]: I1001 09:38:50.114294 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7p4xb" podUID="52a18549-2c1c-4f28-bfa1-6fb14e0690e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 01 09:38:50 crc kubenswrapper[4787]: I1001 09:38:50.514838 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0eb4a902-9efa-4088-a236-9a634fd085fd-metrics-certs\") pod \"network-metrics-daemon-6zz6g\" (UID: \"0eb4a902-9efa-4088-a236-9a634fd085fd\") " pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:38:50 crc kubenswrapper[4787]: I1001 09:38:50.528220 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0eb4a902-9efa-4088-a236-9a634fd085fd-metrics-certs\") pod \"network-metrics-daemon-6zz6g\" (UID: \"0eb4a902-9efa-4088-a236-9a634fd085fd\") " pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:38:50 crc kubenswrapper[4787]: I1001 09:38:50.661713 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6zz6g" Oct 01 09:38:51 crc kubenswrapper[4787]: I1001 09:38:51.224995 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-7p4xb" Oct 01 09:38:51 crc kubenswrapper[4787]: I1001 09:38:51.228823 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-7p4xb" Oct 01 09:38:56 crc kubenswrapper[4787]: I1001 09:38:56.053705 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 09:38:56 crc kubenswrapper[4787]: I1001 09:38:56.202907 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9ec992e3-443f-4601-9faf-ca18c832a8db-kubelet-dir\") pod \"9ec992e3-443f-4601-9faf-ca18c832a8db\" (UID: \"9ec992e3-443f-4601-9faf-ca18c832a8db\") " Oct 01 09:38:56 crc kubenswrapper[4787]: I1001 09:38:56.203012 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9ec992e3-443f-4601-9faf-ca18c832a8db-kube-api-access\") pod \"9ec992e3-443f-4601-9faf-ca18c832a8db\" (UID: \"9ec992e3-443f-4601-9faf-ca18c832a8db\") " Oct 01 09:38:56 crc kubenswrapper[4787]: I1001 09:38:56.203116 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ec992e3-443f-4601-9faf-ca18c832a8db-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "9ec992e3-443f-4601-9faf-ca18c832a8db" (UID: "9ec992e3-443f-4601-9faf-ca18c832a8db"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:38:56 crc kubenswrapper[4787]: I1001 09:38:56.203424 4787 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9ec992e3-443f-4601-9faf-ca18c832a8db-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 01 09:38:56 crc kubenswrapper[4787]: I1001 09:38:56.217266 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ec992e3-443f-4601-9faf-ca18c832a8db-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "9ec992e3-443f-4601-9faf-ca18c832a8db" (UID: "9ec992e3-443f-4601-9faf-ca18c832a8db"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:38:56 crc kubenswrapper[4787]: I1001 09:38:56.304647 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9ec992e3-443f-4601-9faf-ca18c832a8db-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 01 09:38:56 crc kubenswrapper[4787]: I1001 09:38:56.540428 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9ec992e3-443f-4601-9faf-ca18c832a8db","Type":"ContainerDied","Data":"b0e60a24fbbeab813f76e771c903db9d4b46a67e4fb43b8ebb74a76d6fb94660"} Oct 01 09:38:56 crc kubenswrapper[4787]: I1001 09:38:56.540494 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b0e60a24fbbeab813f76e771c903db9d4b46a67e4fb43b8ebb74a76d6fb94660" Oct 01 09:38:56 crc kubenswrapper[4787]: I1001 09:38:56.540610 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 01 09:38:58 crc kubenswrapper[4787]: I1001 09:38:58.075718 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:38:59 crc kubenswrapper[4787]: I1001 09:38:59.927409 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-gqvt9" Oct 01 09:38:59 crc kubenswrapper[4787]: I1001 09:38:59.930958 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-gqvt9" Oct 01 09:39:04 crc kubenswrapper[4787]: E1001 09:39:04.377633 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 01 09:39:04 crc kubenswrapper[4787]: E1001 09:39:04.378421 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s7fht,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-cqr6k_openshift-marketplace(c4fd105d-9feb-4898-92a1-9c7ae1613202): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 01 09:39:04 crc kubenswrapper[4787]: E1001 09:39:04.379624 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-cqr6k" podUID="c4fd105d-9feb-4898-92a1-9c7ae1613202" Oct 01 09:39:07 crc kubenswrapper[4787]: E1001 09:39:07.196065 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-cqr6k" podUID="c4fd105d-9feb-4898-92a1-9c7ae1613202" Oct 01 09:39:08 crc kubenswrapper[4787]: E1001 09:39:08.486721 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 01 09:39:08 crc kubenswrapper[4787]: E1001 09:39:08.486890 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gpzd6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-xq2kw_openshift-marketplace(045fb03e-9b60-4729-b4cd-73db79bb6294): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 01 09:39:08 crc kubenswrapper[4787]: E1001 09:39:08.488687 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-xq2kw" podUID="045fb03e-9b60-4729-b4cd-73db79bb6294" Oct 01 09:39:09 crc kubenswrapper[4787]: E1001 09:39:09.157876 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xq2kw" podUID="045fb03e-9b60-4729-b4cd-73db79bb6294" Oct 01 09:39:09 crc kubenswrapper[4787]: E1001 09:39:09.233036 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 01 09:39:09 crc kubenswrapper[4787]: E1001 09:39:09.233412 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jvxmq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-tzk9m_openshift-marketplace(aa6719df-be6a-4209-9f40-70fbba0465a1): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 01 09:39:09 crc kubenswrapper[4787]: E1001 09:39:09.234638 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-tzk9m" podUID="aa6719df-be6a-4209-9f40-70fbba0465a1" Oct 01 09:39:09 crc kubenswrapper[4787]: E1001 09:39:09.346153 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 01 09:39:09 crc kubenswrapper[4787]: E1001 09:39:09.346468 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k587l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-jszln_openshift-marketplace(34c3f769-241a-4df9-9f8d-48308997f85c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 01 09:39:09 crc kubenswrapper[4787]: E1001 09:39:09.347693 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-jszln" podUID="34c3f769-241a-4df9-9f8d-48308997f85c" Oct 01 09:39:09 crc kubenswrapper[4787]: I1001 09:39:09.627541 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-6zz6g"] Oct 01 09:39:09 crc kubenswrapper[4787]: I1001 09:39:09.632282 4787 generic.go:334] "Generic (PLEG): container finished" podID="315c2b9e-c2cc-4192-8a74-fe35860ff463" containerID="e69780c2e3002e5e68b1853ff6e3ce99d978d5cbaf6c1a5da0bd01924665f584" exitCode=0 Oct 01 09:39:09 crc kubenswrapper[4787]: I1001 09:39:09.632416 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z2fsb" event={"ID":"315c2b9e-c2cc-4192-8a74-fe35860ff463","Type":"ContainerDied","Data":"e69780c2e3002e5e68b1853ff6e3ce99d978d5cbaf6c1a5da0bd01924665f584"} Oct 01 09:39:09 crc kubenswrapper[4787]: I1001 09:39:09.633970 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n72ph" event={"ID":"0637215a-7749-4b92-a60e-f24e2d1b1a29","Type":"ContainerStarted","Data":"affd09ec48ca0a19d1cffb1338b4e450ed836e75b2a18306e8c67f7ce7c82b66"} Oct 01 09:39:09 crc kubenswrapper[4787]: W1001 09:39:09.636380 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0eb4a902_9efa_4088_a236_9a634fd085fd.slice/crio-2c26e0cdb3ef30a8962be5bcfa6b67d29eaf27ccf3f2830d08dff3c934bba2b9 WatchSource:0}: Error finding container 2c26e0cdb3ef30a8962be5bcfa6b67d29eaf27ccf3f2830d08dff3c934bba2b9: Status 404 returned error can't find the container with id 2c26e0cdb3ef30a8962be5bcfa6b67d29eaf27ccf3f2830d08dff3c934bba2b9 Oct 01 09:39:09 crc kubenswrapper[4787]: I1001 09:39:09.640549 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t6rc9" event={"ID":"4914dc93-bedf-4429-a399-9f465e2c0ab9","Type":"ContainerStarted","Data":"165cb69676439a25be6b8ad9e9a51e2d18e2a76c427cb8b2f3e278f638b28260"} Oct 01 09:39:09 crc kubenswrapper[4787]: I1001 09:39:09.647504 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5759" event={"ID":"33bdb919-2c9f-4133-be8b-368dea5b099a","Type":"ContainerStarted","Data":"e900e6e6cf30fdc8da1392ab8278535201226dc781c6769047ad6be7a79375f5"} Oct 01 09:39:09 crc kubenswrapper[4787]: E1001 09:39:09.654421 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-jszln" podUID="34c3f769-241a-4df9-9f8d-48308997f85c" Oct 01 09:39:09 crc kubenswrapper[4787]: E1001 09:39:09.657902 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-tzk9m" podUID="aa6719df-be6a-4209-9f40-70fbba0465a1" Oct 01 09:39:10 crc kubenswrapper[4787]: I1001 09:39:10.227930 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-blwqt" Oct 01 09:39:10 crc kubenswrapper[4787]: I1001 09:39:10.663218 4787 generic.go:334] "Generic (PLEG): container finished" podID="0637215a-7749-4b92-a60e-f24e2d1b1a29" containerID="affd09ec48ca0a19d1cffb1338b4e450ed836e75b2a18306e8c67f7ce7c82b66" exitCode=0 Oct 01 09:39:10 crc kubenswrapper[4787]: I1001 09:39:10.663347 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n72ph" event={"ID":"0637215a-7749-4b92-a60e-f24e2d1b1a29","Type":"ContainerDied","Data":"affd09ec48ca0a19d1cffb1338b4e450ed836e75b2a18306e8c67f7ce7c82b66"} Oct 01 09:39:10 crc kubenswrapper[4787]: I1001 09:39:10.671393 4787 generic.go:334] "Generic (PLEG): container finished" podID="4914dc93-bedf-4429-a399-9f465e2c0ab9" containerID="165cb69676439a25be6b8ad9e9a51e2d18e2a76c427cb8b2f3e278f638b28260" exitCode=0 Oct 01 09:39:10 crc kubenswrapper[4787]: I1001 09:39:10.671512 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t6rc9" event={"ID":"4914dc93-bedf-4429-a399-9f465e2c0ab9","Type":"ContainerDied","Data":"165cb69676439a25be6b8ad9e9a51e2d18e2a76c427cb8b2f3e278f638b28260"} Oct 01 09:39:10 crc kubenswrapper[4787]: I1001 09:39:10.678857 4787 generic.go:334] "Generic (PLEG): container finished" podID="33bdb919-2c9f-4133-be8b-368dea5b099a" containerID="e900e6e6cf30fdc8da1392ab8278535201226dc781c6769047ad6be7a79375f5" exitCode=0 Oct 01 09:39:10 crc kubenswrapper[4787]: I1001 09:39:10.678953 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5759" event={"ID":"33bdb919-2c9f-4133-be8b-368dea5b099a","Type":"ContainerDied","Data":"e900e6e6cf30fdc8da1392ab8278535201226dc781c6769047ad6be7a79375f5"} Oct 01 09:39:10 crc kubenswrapper[4787]: I1001 09:39:10.685865 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-6zz6g" event={"ID":"0eb4a902-9efa-4088-a236-9a634fd085fd","Type":"ContainerStarted","Data":"ae406f2f7ea29c57e914d80325c04059e670ca9a2589a0e04dc99674f8514050"} Oct 01 09:39:10 crc kubenswrapper[4787]: I1001 09:39:10.685927 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-6zz6g" event={"ID":"0eb4a902-9efa-4088-a236-9a634fd085fd","Type":"ContainerStarted","Data":"2c26e0cdb3ef30a8962be5bcfa6b67d29eaf27ccf3f2830d08dff3c934bba2b9"} Oct 01 09:39:11 crc kubenswrapper[4787]: I1001 09:39:11.250809 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:39:11 crc kubenswrapper[4787]: I1001 09:39:11.250877 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:39:11 crc kubenswrapper[4787]: I1001 09:39:11.696498 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-6zz6g" event={"ID":"0eb4a902-9efa-4088-a236-9a634fd085fd","Type":"ContainerStarted","Data":"b42f093d24fdd733ed662d1cddadb00a40c4fe1e9e11a3d1026fd30ed956e7b1"} Oct 01 09:39:11 crc kubenswrapper[4787]: I1001 09:39:11.735903 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-6zz6g" podStartSLOduration=163.735872782 podStartE2EDuration="2m43.735872782s" podCreationTimestamp="2025-10-01 09:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:39:11.715633563 +0000 UTC m=+183.830777720" watchObservedRunningTime="2025-10-01 09:39:11.735872782 +0000 UTC m=+183.851016939" Oct 01 09:39:12 crc kubenswrapper[4787]: I1001 09:39:12.704864 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t6rc9" event={"ID":"4914dc93-bedf-4429-a399-9f465e2c0ab9","Type":"ContainerStarted","Data":"18558e35e002ca21b2ec847e1bb25fb54f8a3f5605ad161eeb2e007c88f96a22"} Oct 01 09:39:12 crc kubenswrapper[4787]: I1001 09:39:12.709315 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5759" event={"ID":"33bdb919-2c9f-4133-be8b-368dea5b099a","Type":"ContainerStarted","Data":"72ec9ac17e4f19189400cc7b371235a0844c21660dcc5604203f2c47654b02ff"} Oct 01 09:39:12 crc kubenswrapper[4787]: I1001 09:39:12.713181 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z2fsb" event={"ID":"315c2b9e-c2cc-4192-8a74-fe35860ff463","Type":"ContainerStarted","Data":"ba3bfd8b91dad1bc5fbd89dc0c8b5bea4a527995ca6d733a5394a85f088cd3e6"} Oct 01 09:39:12 crc kubenswrapper[4787]: I1001 09:39:12.716948 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n72ph" event={"ID":"0637215a-7749-4b92-a60e-f24e2d1b1a29","Type":"ContainerStarted","Data":"e01bf7b28841598cb7bbefaa663d8849342b23617d5f0282ab457c1ae89f472a"} Oct 01 09:39:12 crc kubenswrapper[4787]: I1001 09:39:12.748390 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-t6rc9" podStartSLOduration=2.6045793760000002 podStartE2EDuration="34.748363881s" podCreationTimestamp="2025-10-01 09:38:38 +0000 UTC" firstStartedPulling="2025-10-01 09:38:40.233588679 +0000 UTC m=+152.348732836" lastFinishedPulling="2025-10-01 09:39:12.377373174 +0000 UTC m=+184.492517341" observedRunningTime="2025-10-01 09:39:12.7319139 +0000 UTC m=+184.847058077" watchObservedRunningTime="2025-10-01 09:39:12.748363881 +0000 UTC m=+184.863508038" Oct 01 09:39:12 crc kubenswrapper[4787]: I1001 09:39:12.770713 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-p5759" podStartSLOduration=3.784673667 podStartE2EDuration="34.770687523s" podCreationTimestamp="2025-10-01 09:38:38 +0000 UTC" firstStartedPulling="2025-10-01 09:38:41.282252665 +0000 UTC m=+153.397396822" lastFinishedPulling="2025-10-01 09:39:12.268266521 +0000 UTC m=+184.383410678" observedRunningTime="2025-10-01 09:39:12.770242691 +0000 UTC m=+184.885386858" watchObservedRunningTime="2025-10-01 09:39:12.770687523 +0000 UTC m=+184.885831680" Oct 01 09:39:12 crc kubenswrapper[4787]: I1001 09:39:12.774510 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-z2fsb" podStartSLOduration=2.848384439 podStartE2EDuration="35.77449292s" podCreationTimestamp="2025-10-01 09:38:37 +0000 UTC" firstStartedPulling="2025-10-01 09:38:39.222235369 +0000 UTC m=+151.337379526" lastFinishedPulling="2025-10-01 09:39:12.14834383 +0000 UTC m=+184.263488007" observedRunningTime="2025-10-01 09:39:12.751285606 +0000 UTC m=+184.866429783" watchObservedRunningTime="2025-10-01 09:39:12.77449292 +0000 UTC m=+184.889637077" Oct 01 09:39:12 crc kubenswrapper[4787]: I1001 09:39:12.795686 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-n72ph" podStartSLOduration=3.709903213 podStartE2EDuration="37.795665723s" podCreationTimestamp="2025-10-01 09:38:35 +0000 UTC" firstStartedPulling="2025-10-01 09:38:38.103989421 +0000 UTC m=+150.219133578" lastFinishedPulling="2025-10-01 09:39:12.189751931 +0000 UTC m=+184.304896088" observedRunningTime="2025-10-01 09:39:12.794314698 +0000 UTC m=+184.909458875" watchObservedRunningTime="2025-10-01 09:39:12.795665723 +0000 UTC m=+184.910809880" Oct 01 09:39:16 crc kubenswrapper[4787]: I1001 09:39:16.011042 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-n72ph" Oct 01 09:39:16 crc kubenswrapper[4787]: I1001 09:39:16.011653 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-n72ph" Oct 01 09:39:16 crc kubenswrapper[4787]: I1001 09:39:16.141053 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-n72ph" Oct 01 09:39:16 crc kubenswrapper[4787]: I1001 09:39:16.654663 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 01 09:39:17 crc kubenswrapper[4787]: I1001 09:39:17.733134 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-z2fsb" Oct 01 09:39:17 crc kubenswrapper[4787]: I1001 09:39:17.733246 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-z2fsb" Oct 01 09:39:17 crc kubenswrapper[4787]: I1001 09:39:17.791065 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-z2fsb" Oct 01 09:39:17 crc kubenswrapper[4787]: I1001 09:39:17.829951 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-z2fsb" Oct 01 09:39:18 crc kubenswrapper[4787]: I1001 09:39:18.941427 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-t6rc9" Oct 01 09:39:18 crc kubenswrapper[4787]: I1001 09:39:18.942163 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-t6rc9" Oct 01 09:39:18 crc kubenswrapper[4787]: I1001 09:39:18.987637 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-t6rc9" Oct 01 09:39:19 crc kubenswrapper[4787]: I1001 09:39:19.326948 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-p5759" Oct 01 09:39:19 crc kubenswrapper[4787]: I1001 09:39:19.327001 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-p5759" Oct 01 09:39:19 crc kubenswrapper[4787]: I1001 09:39:19.387534 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-p5759" Oct 01 09:39:19 crc kubenswrapper[4787]: I1001 09:39:19.796918 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-p5759" Oct 01 09:39:19 crc kubenswrapper[4787]: I1001 09:39:19.800228 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-t6rc9" Oct 01 09:39:20 crc kubenswrapper[4787]: I1001 09:39:20.857035 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p5759"] Oct 01 09:39:21 crc kubenswrapper[4787]: I1001 09:39:21.764421 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-p5759" podUID="33bdb919-2c9f-4133-be8b-368dea5b099a" containerName="registry-server" containerID="cri-o://72ec9ac17e4f19189400cc7b371235a0844c21660dcc5604203f2c47654b02ff" gracePeriod=2 Oct 01 09:39:22 crc kubenswrapper[4787]: I1001 09:39:22.783601 4787 generic.go:334] "Generic (PLEG): container finished" podID="33bdb919-2c9f-4133-be8b-368dea5b099a" containerID="72ec9ac17e4f19189400cc7b371235a0844c21660dcc5604203f2c47654b02ff" exitCode=0 Oct 01 09:39:22 crc kubenswrapper[4787]: I1001 09:39:22.783834 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5759" event={"ID":"33bdb919-2c9f-4133-be8b-368dea5b099a","Type":"ContainerDied","Data":"72ec9ac17e4f19189400cc7b371235a0844c21660dcc5604203f2c47654b02ff"} Oct 01 09:39:25 crc kubenswrapper[4787]: I1001 09:39:25.280986 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p5759" Oct 01 09:39:25 crc kubenswrapper[4787]: I1001 09:39:25.443230 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33bdb919-2c9f-4133-be8b-368dea5b099a-utilities\") pod \"33bdb919-2c9f-4133-be8b-368dea5b099a\" (UID: \"33bdb919-2c9f-4133-be8b-368dea5b099a\") " Oct 01 09:39:25 crc kubenswrapper[4787]: I1001 09:39:25.443306 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33bdb919-2c9f-4133-be8b-368dea5b099a-catalog-content\") pod \"33bdb919-2c9f-4133-be8b-368dea5b099a\" (UID: \"33bdb919-2c9f-4133-be8b-368dea5b099a\") " Oct 01 09:39:25 crc kubenswrapper[4787]: I1001 09:39:25.443339 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8hlk\" (UniqueName: \"kubernetes.io/projected/33bdb919-2c9f-4133-be8b-368dea5b099a-kube-api-access-g8hlk\") pod \"33bdb919-2c9f-4133-be8b-368dea5b099a\" (UID: \"33bdb919-2c9f-4133-be8b-368dea5b099a\") " Oct 01 09:39:25 crc kubenswrapper[4787]: I1001 09:39:25.444191 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33bdb919-2c9f-4133-be8b-368dea5b099a-utilities" (OuterVolumeSpecName: "utilities") pod "33bdb919-2c9f-4133-be8b-368dea5b099a" (UID: "33bdb919-2c9f-4133-be8b-368dea5b099a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:39:25 crc kubenswrapper[4787]: I1001 09:39:25.449142 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33bdb919-2c9f-4133-be8b-368dea5b099a-kube-api-access-g8hlk" (OuterVolumeSpecName: "kube-api-access-g8hlk") pod "33bdb919-2c9f-4133-be8b-368dea5b099a" (UID: "33bdb919-2c9f-4133-be8b-368dea5b099a"). InnerVolumeSpecName "kube-api-access-g8hlk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:39:25 crc kubenswrapper[4787]: I1001 09:39:25.522009 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33bdb919-2c9f-4133-be8b-368dea5b099a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "33bdb919-2c9f-4133-be8b-368dea5b099a" (UID: "33bdb919-2c9f-4133-be8b-368dea5b099a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:39:25 crc kubenswrapper[4787]: I1001 09:39:25.544612 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33bdb919-2c9f-4133-be8b-368dea5b099a-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:39:25 crc kubenswrapper[4787]: I1001 09:39:25.544641 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33bdb919-2c9f-4133-be8b-368dea5b099a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:39:25 crc kubenswrapper[4787]: I1001 09:39:25.544652 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8hlk\" (UniqueName: \"kubernetes.io/projected/33bdb919-2c9f-4133-be8b-368dea5b099a-kube-api-access-g8hlk\") on node \"crc\" DevicePath \"\"" Oct 01 09:39:25 crc kubenswrapper[4787]: I1001 09:39:25.810758 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5759" event={"ID":"33bdb919-2c9f-4133-be8b-368dea5b099a","Type":"ContainerDied","Data":"d96d47b1ab4af429e19124794a21315419157ddbd78f5109f7e42c6c2ce4d4d9"} Oct 01 09:39:25 crc kubenswrapper[4787]: I1001 09:39:25.810829 4787 scope.go:117] "RemoveContainer" containerID="72ec9ac17e4f19189400cc7b371235a0844c21660dcc5604203f2c47654b02ff" Oct 01 09:39:25 crc kubenswrapper[4787]: I1001 09:39:25.810797 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p5759" Oct 01 09:39:25 crc kubenswrapper[4787]: I1001 09:39:25.837938 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p5759"] Oct 01 09:39:25 crc kubenswrapper[4787]: I1001 09:39:25.843269 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-p5759"] Oct 01 09:39:26 crc kubenswrapper[4787]: I1001 09:39:26.054563 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-n72ph" Oct 01 09:39:26 crc kubenswrapper[4787]: I1001 09:39:26.363888 4787 scope.go:117] "RemoveContainer" containerID="e900e6e6cf30fdc8da1392ab8278535201226dc781c6769047ad6be7a79375f5" Oct 01 09:39:26 crc kubenswrapper[4787]: I1001 09:39:26.385526 4787 scope.go:117] "RemoveContainer" containerID="9737f3591d2c84e57b5e6fd3b421a8b9ac9aaa395227e2da1442a4ff1c592a6c" Oct 01 09:39:26 crc kubenswrapper[4787]: I1001 09:39:26.529882 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33bdb919-2c9f-4133-be8b-368dea5b099a" path="/var/lib/kubelet/pods/33bdb919-2c9f-4133-be8b-368dea5b099a/volumes" Oct 01 09:39:27 crc kubenswrapper[4787]: I1001 09:39:27.260880 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-n72ph"] Oct 01 09:39:27 crc kubenswrapper[4787]: I1001 09:39:27.261268 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-n72ph" podUID="0637215a-7749-4b92-a60e-f24e2d1b1a29" containerName="registry-server" containerID="cri-o://e01bf7b28841598cb7bbefaa663d8849342b23617d5f0282ab457c1ae89f472a" gracePeriod=2 Oct 01 09:39:27 crc kubenswrapper[4787]: I1001 09:39:27.640037 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n72ph" Oct 01 09:39:27 crc kubenswrapper[4787]: I1001 09:39:27.778733 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzx4v\" (UniqueName: \"kubernetes.io/projected/0637215a-7749-4b92-a60e-f24e2d1b1a29-kube-api-access-kzx4v\") pod \"0637215a-7749-4b92-a60e-f24e2d1b1a29\" (UID: \"0637215a-7749-4b92-a60e-f24e2d1b1a29\") " Oct 01 09:39:27 crc kubenswrapper[4787]: I1001 09:39:27.779148 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0637215a-7749-4b92-a60e-f24e2d1b1a29-catalog-content\") pod \"0637215a-7749-4b92-a60e-f24e2d1b1a29\" (UID: \"0637215a-7749-4b92-a60e-f24e2d1b1a29\") " Oct 01 09:39:27 crc kubenswrapper[4787]: I1001 09:39:27.779234 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0637215a-7749-4b92-a60e-f24e2d1b1a29-utilities\") pod \"0637215a-7749-4b92-a60e-f24e2d1b1a29\" (UID: \"0637215a-7749-4b92-a60e-f24e2d1b1a29\") " Oct 01 09:39:27 crc kubenswrapper[4787]: I1001 09:39:27.780375 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0637215a-7749-4b92-a60e-f24e2d1b1a29-utilities" (OuterVolumeSpecName: "utilities") pod "0637215a-7749-4b92-a60e-f24e2d1b1a29" (UID: "0637215a-7749-4b92-a60e-f24e2d1b1a29"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:39:27 crc kubenswrapper[4787]: I1001 09:39:27.797688 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0637215a-7749-4b92-a60e-f24e2d1b1a29-kube-api-access-kzx4v" (OuterVolumeSpecName: "kube-api-access-kzx4v") pod "0637215a-7749-4b92-a60e-f24e2d1b1a29" (UID: "0637215a-7749-4b92-a60e-f24e2d1b1a29"). InnerVolumeSpecName "kube-api-access-kzx4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:39:27 crc kubenswrapper[4787]: I1001 09:39:27.830044 4787 generic.go:334] "Generic (PLEG): container finished" podID="045fb03e-9b60-4729-b4cd-73db79bb6294" containerID="970bc197a3226fc571cfc9c8e6c8725e78105d5112b1bcb73de8f125d6a62801" exitCode=0 Oct 01 09:39:27 crc kubenswrapper[4787]: I1001 09:39:27.830120 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xq2kw" event={"ID":"045fb03e-9b60-4729-b4cd-73db79bb6294","Type":"ContainerDied","Data":"970bc197a3226fc571cfc9c8e6c8725e78105d5112b1bcb73de8f125d6a62801"} Oct 01 09:39:27 crc kubenswrapper[4787]: I1001 09:39:27.832307 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0637215a-7749-4b92-a60e-f24e2d1b1a29-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0637215a-7749-4b92-a60e-f24e2d1b1a29" (UID: "0637215a-7749-4b92-a60e-f24e2d1b1a29"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:39:27 crc kubenswrapper[4787]: I1001 09:39:27.832781 4787 generic.go:334] "Generic (PLEG): container finished" podID="0637215a-7749-4b92-a60e-f24e2d1b1a29" containerID="e01bf7b28841598cb7bbefaa663d8849342b23617d5f0282ab457c1ae89f472a" exitCode=0 Oct 01 09:39:27 crc kubenswrapper[4787]: I1001 09:39:27.832807 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n72ph" event={"ID":"0637215a-7749-4b92-a60e-f24e2d1b1a29","Type":"ContainerDied","Data":"e01bf7b28841598cb7bbefaa663d8849342b23617d5f0282ab457c1ae89f472a"} Oct 01 09:39:27 crc kubenswrapper[4787]: I1001 09:39:27.832824 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n72ph" event={"ID":"0637215a-7749-4b92-a60e-f24e2d1b1a29","Type":"ContainerDied","Data":"4ff46a5fc2376146771db7fb30bfeb45d340acf1aaf46139a7c1124624387a96"} Oct 01 09:39:27 crc kubenswrapper[4787]: I1001 09:39:27.832840 4787 scope.go:117] "RemoveContainer" containerID="e01bf7b28841598cb7bbefaa663d8849342b23617d5f0282ab457c1ae89f472a" Oct 01 09:39:27 crc kubenswrapper[4787]: I1001 09:39:27.832939 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n72ph" Oct 01 09:39:27 crc kubenswrapper[4787]: I1001 09:39:27.869270 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-n72ph"] Oct 01 09:39:27 crc kubenswrapper[4787]: I1001 09:39:27.872454 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-n72ph"] Oct 01 09:39:27 crc kubenswrapper[4787]: I1001 09:39:27.880423 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzx4v\" (UniqueName: \"kubernetes.io/projected/0637215a-7749-4b92-a60e-f24e2d1b1a29-kube-api-access-kzx4v\") on node \"crc\" DevicePath \"\"" Oct 01 09:39:27 crc kubenswrapper[4787]: I1001 09:39:27.880502 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0637215a-7749-4b92-a60e-f24e2d1b1a29-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:39:27 crc kubenswrapper[4787]: I1001 09:39:27.880582 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0637215a-7749-4b92-a60e-f24e2d1b1a29-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:39:27 crc kubenswrapper[4787]: I1001 09:39:27.944401 4787 scope.go:117] "RemoveContainer" containerID="affd09ec48ca0a19d1cffb1338b4e450ed836e75b2a18306e8c67f7ce7c82b66" Oct 01 09:39:28 crc kubenswrapper[4787]: I1001 09:39:28.283867 4787 scope.go:117] "RemoveContainer" containerID="992bf39e914a8b5f27b7448fd98046bf25404b4039770bca150fbfdf6f29d667" Oct 01 09:39:28 crc kubenswrapper[4787]: I1001 09:39:28.303419 4787 scope.go:117] "RemoveContainer" containerID="e01bf7b28841598cb7bbefaa663d8849342b23617d5f0282ab457c1ae89f472a" Oct 01 09:39:28 crc kubenswrapper[4787]: E1001 09:39:28.303917 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e01bf7b28841598cb7bbefaa663d8849342b23617d5f0282ab457c1ae89f472a\": container with ID starting with e01bf7b28841598cb7bbefaa663d8849342b23617d5f0282ab457c1ae89f472a not found: ID does not exist" containerID="e01bf7b28841598cb7bbefaa663d8849342b23617d5f0282ab457c1ae89f472a" Oct 01 09:39:28 crc kubenswrapper[4787]: I1001 09:39:28.303977 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e01bf7b28841598cb7bbefaa663d8849342b23617d5f0282ab457c1ae89f472a"} err="failed to get container status \"e01bf7b28841598cb7bbefaa663d8849342b23617d5f0282ab457c1ae89f472a\": rpc error: code = NotFound desc = could not find container \"e01bf7b28841598cb7bbefaa663d8849342b23617d5f0282ab457c1ae89f472a\": container with ID starting with e01bf7b28841598cb7bbefaa663d8849342b23617d5f0282ab457c1ae89f472a not found: ID does not exist" Oct 01 09:39:28 crc kubenswrapper[4787]: I1001 09:39:28.304018 4787 scope.go:117] "RemoveContainer" containerID="affd09ec48ca0a19d1cffb1338b4e450ed836e75b2a18306e8c67f7ce7c82b66" Oct 01 09:39:28 crc kubenswrapper[4787]: E1001 09:39:28.304521 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"affd09ec48ca0a19d1cffb1338b4e450ed836e75b2a18306e8c67f7ce7c82b66\": container with ID starting with affd09ec48ca0a19d1cffb1338b4e450ed836e75b2a18306e8c67f7ce7c82b66 not found: ID does not exist" containerID="affd09ec48ca0a19d1cffb1338b4e450ed836e75b2a18306e8c67f7ce7c82b66" Oct 01 09:39:28 crc kubenswrapper[4787]: I1001 09:39:28.304559 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"affd09ec48ca0a19d1cffb1338b4e450ed836e75b2a18306e8c67f7ce7c82b66"} err="failed to get container status \"affd09ec48ca0a19d1cffb1338b4e450ed836e75b2a18306e8c67f7ce7c82b66\": rpc error: code = NotFound desc = could not find container \"affd09ec48ca0a19d1cffb1338b4e450ed836e75b2a18306e8c67f7ce7c82b66\": container with ID starting with affd09ec48ca0a19d1cffb1338b4e450ed836e75b2a18306e8c67f7ce7c82b66 not found: ID does not exist" Oct 01 09:39:28 crc kubenswrapper[4787]: I1001 09:39:28.304585 4787 scope.go:117] "RemoveContainer" containerID="992bf39e914a8b5f27b7448fd98046bf25404b4039770bca150fbfdf6f29d667" Oct 01 09:39:28 crc kubenswrapper[4787]: E1001 09:39:28.305036 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"992bf39e914a8b5f27b7448fd98046bf25404b4039770bca150fbfdf6f29d667\": container with ID starting with 992bf39e914a8b5f27b7448fd98046bf25404b4039770bca150fbfdf6f29d667 not found: ID does not exist" containerID="992bf39e914a8b5f27b7448fd98046bf25404b4039770bca150fbfdf6f29d667" Oct 01 09:39:28 crc kubenswrapper[4787]: I1001 09:39:28.305142 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"992bf39e914a8b5f27b7448fd98046bf25404b4039770bca150fbfdf6f29d667"} err="failed to get container status \"992bf39e914a8b5f27b7448fd98046bf25404b4039770bca150fbfdf6f29d667\": rpc error: code = NotFound desc = could not find container \"992bf39e914a8b5f27b7448fd98046bf25404b4039770bca150fbfdf6f29d667\": container with ID starting with 992bf39e914a8b5f27b7448fd98046bf25404b4039770bca150fbfdf6f29d667 not found: ID does not exist" Oct 01 09:39:28 crc kubenswrapper[4787]: I1001 09:39:28.534603 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0637215a-7749-4b92-a60e-f24e2d1b1a29" path="/var/lib/kubelet/pods/0637215a-7749-4b92-a60e-f24e2d1b1a29/volumes" Oct 01 09:39:28 crc kubenswrapper[4787]: I1001 09:39:28.840758 4787 generic.go:334] "Generic (PLEG): container finished" podID="c4fd105d-9feb-4898-92a1-9c7ae1613202" containerID="939dca2b60d466af76455a0b625d2177e4f2871161a5176c3631d1824a7d7d79" exitCode=0 Oct 01 09:39:28 crc kubenswrapper[4787]: I1001 09:39:28.840817 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cqr6k" event={"ID":"c4fd105d-9feb-4898-92a1-9c7ae1613202","Type":"ContainerDied","Data":"939dca2b60d466af76455a0b625d2177e4f2871161a5176c3631d1824a7d7d79"} Oct 01 09:39:28 crc kubenswrapper[4787]: I1001 09:39:28.848204 4787 generic.go:334] "Generic (PLEG): container finished" podID="aa6719df-be6a-4209-9f40-70fbba0465a1" containerID="2fa9815afd9594d010c212553024c1410df3d8aa855359dcf4bd88ca07d850d3" exitCode=0 Oct 01 09:39:28 crc kubenswrapper[4787]: I1001 09:39:28.848247 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tzk9m" event={"ID":"aa6719df-be6a-4209-9f40-70fbba0465a1","Type":"ContainerDied","Data":"2fa9815afd9594d010c212553024c1410df3d8aa855359dcf4bd88ca07d850d3"} Oct 01 09:39:29 crc kubenswrapper[4787]: I1001 09:39:29.865398 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xq2kw" event={"ID":"045fb03e-9b60-4729-b4cd-73db79bb6294","Type":"ContainerStarted","Data":"6e5aa283bce46500c27c7ce2816d6399083309226b0a0c302494b888e82609ba"} Oct 01 09:39:29 crc kubenswrapper[4787]: I1001 09:39:29.868471 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cqr6k" event={"ID":"c4fd105d-9feb-4898-92a1-9c7ae1613202","Type":"ContainerStarted","Data":"470dad0110c39dc6d9f70ec7a0c4b95126dcc8304888d81e74292727dfeec681"} Oct 01 09:39:29 crc kubenswrapper[4787]: I1001 09:39:29.872405 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tzk9m" event={"ID":"aa6719df-be6a-4209-9f40-70fbba0465a1","Type":"ContainerStarted","Data":"7a07abadeccb594a8a4b687220bc38f001beb285909f3bfa2655ce9f6a7314c8"} Oct 01 09:39:29 crc kubenswrapper[4787]: I1001 09:39:29.875301 4787 generic.go:334] "Generic (PLEG): container finished" podID="34c3f769-241a-4df9-9f8d-48308997f85c" containerID="e17a2054049276967a4ee7888f5f986176669408fb2c1fc161f9e0d338de2c68" exitCode=0 Oct 01 09:39:29 crc kubenswrapper[4787]: I1001 09:39:29.875377 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jszln" event={"ID":"34c3f769-241a-4df9-9f8d-48308997f85c","Type":"ContainerDied","Data":"e17a2054049276967a4ee7888f5f986176669408fb2c1fc161f9e0d338de2c68"} Oct 01 09:39:29 crc kubenswrapper[4787]: I1001 09:39:29.886119 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xq2kw" podStartSLOduration=4.066235255 podStartE2EDuration="54.886098043s" podCreationTimestamp="2025-10-01 09:38:35 +0000 UTC" firstStartedPulling="2025-10-01 09:38:38.161321269 +0000 UTC m=+150.276465426" lastFinishedPulling="2025-10-01 09:39:28.981184057 +0000 UTC m=+201.096328214" observedRunningTime="2025-10-01 09:39:29.883416242 +0000 UTC m=+201.998560409" watchObservedRunningTime="2025-10-01 09:39:29.886098043 +0000 UTC m=+202.001242200" Oct 01 09:39:29 crc kubenswrapper[4787]: I1001 09:39:29.919659 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cqr6k" podStartSLOduration=3.803478468 podStartE2EDuration="54.919636902s" podCreationTimestamp="2025-10-01 09:38:35 +0000 UTC" firstStartedPulling="2025-10-01 09:38:38.104429343 +0000 UTC m=+150.219573500" lastFinishedPulling="2025-10-01 09:39:29.220587777 +0000 UTC m=+201.335731934" observedRunningTime="2025-10-01 09:39:29.917508685 +0000 UTC m=+202.032652862" watchObservedRunningTime="2025-10-01 09:39:29.919636902 +0000 UTC m=+202.034781059" Oct 01 09:39:29 crc kubenswrapper[4787]: I1001 09:39:29.954042 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tzk9m" podStartSLOduration=2.839578558 podStartE2EDuration="52.954022782s" podCreationTimestamp="2025-10-01 09:38:37 +0000 UTC" firstStartedPulling="2025-10-01 09:38:39.189497811 +0000 UTC m=+151.304641968" lastFinishedPulling="2025-10-01 09:39:29.303942035 +0000 UTC m=+201.419086192" observedRunningTime="2025-10-01 09:39:29.953905629 +0000 UTC m=+202.069049786" watchObservedRunningTime="2025-10-01 09:39:29.954022782 +0000 UTC m=+202.069166939" Oct 01 09:39:30 crc kubenswrapper[4787]: I1001 09:39:30.882158 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jszln" event={"ID":"34c3f769-241a-4df9-9f8d-48308997f85c","Type":"ContainerStarted","Data":"f9ee46727fb7e8eb6dec12809885a91b1ecd1938a84356c816455aa177174f31"} Oct 01 09:39:30 crc kubenswrapper[4787]: I1001 09:39:30.920511 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jszln" podStartSLOduration=3.530030567 podStartE2EDuration="55.920478488s" podCreationTimestamp="2025-10-01 09:38:35 +0000 UTC" firstStartedPulling="2025-10-01 09:38:38.16996945 +0000 UTC m=+150.285113607" lastFinishedPulling="2025-10-01 09:39:30.560417371 +0000 UTC m=+202.675561528" observedRunningTime="2025-10-01 09:39:30.904993077 +0000 UTC m=+203.020137234" watchObservedRunningTime="2025-10-01 09:39:30.920478488 +0000 UTC m=+203.035622645" Oct 01 09:39:33 crc kubenswrapper[4787]: I1001 09:39:33.823003 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-hwftw"] Oct 01 09:39:35 crc kubenswrapper[4787]: I1001 09:39:35.575179 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cqr6k" Oct 01 09:39:35 crc kubenswrapper[4787]: I1001 09:39:35.575630 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cqr6k" Oct 01 09:39:35 crc kubenswrapper[4787]: I1001 09:39:35.649583 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cqr6k" Oct 01 09:39:35 crc kubenswrapper[4787]: I1001 09:39:35.738824 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xq2kw" Oct 01 09:39:35 crc kubenswrapper[4787]: I1001 09:39:35.738933 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xq2kw" Oct 01 09:39:35 crc kubenswrapper[4787]: I1001 09:39:35.795381 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xq2kw" Oct 01 09:39:35 crc kubenswrapper[4787]: I1001 09:39:35.958545 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xq2kw" Oct 01 09:39:35 crc kubenswrapper[4787]: I1001 09:39:35.960858 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cqr6k" Oct 01 09:39:36 crc kubenswrapper[4787]: I1001 09:39:36.283466 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jszln" Oct 01 09:39:36 crc kubenswrapper[4787]: I1001 09:39:36.283523 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jszln" Oct 01 09:39:36 crc kubenswrapper[4787]: I1001 09:39:36.328317 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jszln" Oct 01 09:39:36 crc kubenswrapper[4787]: I1001 09:39:36.957873 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jszln" Oct 01 09:39:37 crc kubenswrapper[4787]: I1001 09:39:37.921723 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tzk9m" Oct 01 09:39:37 crc kubenswrapper[4787]: I1001 09:39:37.921768 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tzk9m" Oct 01 09:39:37 crc kubenswrapper[4787]: I1001 09:39:37.958939 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tzk9m" Oct 01 09:39:38 crc kubenswrapper[4787]: I1001 09:39:38.058554 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jszln"] Oct 01 09:39:38 crc kubenswrapper[4787]: I1001 09:39:38.926627 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jszln" podUID="34c3f769-241a-4df9-9f8d-48308997f85c" containerName="registry-server" containerID="cri-o://f9ee46727fb7e8eb6dec12809885a91b1ecd1938a84356c816455aa177174f31" gracePeriod=2 Oct 01 09:39:39 crc kubenswrapper[4787]: I1001 09:39:39.059357 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tzk9m" Oct 01 09:39:39 crc kubenswrapper[4787]: I1001 09:39:39.330039 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jszln" Oct 01 09:39:39 crc kubenswrapper[4787]: I1001 09:39:39.434485 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34c3f769-241a-4df9-9f8d-48308997f85c-utilities\") pod \"34c3f769-241a-4df9-9f8d-48308997f85c\" (UID: \"34c3f769-241a-4df9-9f8d-48308997f85c\") " Oct 01 09:39:39 crc kubenswrapper[4787]: I1001 09:39:39.434669 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34c3f769-241a-4df9-9f8d-48308997f85c-catalog-content\") pod \"34c3f769-241a-4df9-9f8d-48308997f85c\" (UID: \"34c3f769-241a-4df9-9f8d-48308997f85c\") " Oct 01 09:39:39 crc kubenswrapper[4787]: I1001 09:39:39.434722 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k587l\" (UniqueName: \"kubernetes.io/projected/34c3f769-241a-4df9-9f8d-48308997f85c-kube-api-access-k587l\") pod \"34c3f769-241a-4df9-9f8d-48308997f85c\" (UID: \"34c3f769-241a-4df9-9f8d-48308997f85c\") " Oct 01 09:39:39 crc kubenswrapper[4787]: I1001 09:39:39.435477 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34c3f769-241a-4df9-9f8d-48308997f85c-utilities" (OuterVolumeSpecName: "utilities") pod "34c3f769-241a-4df9-9f8d-48308997f85c" (UID: "34c3f769-241a-4df9-9f8d-48308997f85c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:39:39 crc kubenswrapper[4787]: I1001 09:39:39.441028 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34c3f769-241a-4df9-9f8d-48308997f85c-kube-api-access-k587l" (OuterVolumeSpecName: "kube-api-access-k587l") pod "34c3f769-241a-4df9-9f8d-48308997f85c" (UID: "34c3f769-241a-4df9-9f8d-48308997f85c"). InnerVolumeSpecName "kube-api-access-k587l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:39:39 crc kubenswrapper[4787]: I1001 09:39:39.497725 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34c3f769-241a-4df9-9f8d-48308997f85c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "34c3f769-241a-4df9-9f8d-48308997f85c" (UID: "34c3f769-241a-4df9-9f8d-48308997f85c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:39:39 crc kubenswrapper[4787]: I1001 09:39:39.536272 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34c3f769-241a-4df9-9f8d-48308997f85c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:39:39 crc kubenswrapper[4787]: I1001 09:39:39.536328 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k587l\" (UniqueName: \"kubernetes.io/projected/34c3f769-241a-4df9-9f8d-48308997f85c-kube-api-access-k587l\") on node \"crc\" DevicePath \"\"" Oct 01 09:39:39 crc kubenswrapper[4787]: I1001 09:39:39.536347 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34c3f769-241a-4df9-9f8d-48308997f85c-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:39:39 crc kubenswrapper[4787]: I1001 09:39:39.932724 4787 generic.go:334] "Generic (PLEG): container finished" podID="34c3f769-241a-4df9-9f8d-48308997f85c" containerID="f9ee46727fb7e8eb6dec12809885a91b1ecd1938a84356c816455aa177174f31" exitCode=0 Oct 01 09:39:39 crc kubenswrapper[4787]: I1001 09:39:39.932792 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jszln" Oct 01 09:39:39 crc kubenswrapper[4787]: I1001 09:39:39.932796 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jszln" event={"ID":"34c3f769-241a-4df9-9f8d-48308997f85c","Type":"ContainerDied","Data":"f9ee46727fb7e8eb6dec12809885a91b1ecd1938a84356c816455aa177174f31"} Oct 01 09:39:39 crc kubenswrapper[4787]: I1001 09:39:39.932916 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jszln" event={"ID":"34c3f769-241a-4df9-9f8d-48308997f85c","Type":"ContainerDied","Data":"47ee8912dc6b6ba20c6471d3c149f8c238814cdb20bcefa278fb9f500565273d"} Oct 01 09:39:39 crc kubenswrapper[4787]: I1001 09:39:39.932939 4787 scope.go:117] "RemoveContainer" containerID="f9ee46727fb7e8eb6dec12809885a91b1ecd1938a84356c816455aa177174f31" Oct 01 09:39:39 crc kubenswrapper[4787]: I1001 09:39:39.954112 4787 scope.go:117] "RemoveContainer" containerID="e17a2054049276967a4ee7888f5f986176669408fb2c1fc161f9e0d338de2c68" Oct 01 09:39:39 crc kubenswrapper[4787]: I1001 09:39:39.972445 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jszln"] Oct 01 09:39:39 crc kubenswrapper[4787]: I1001 09:39:39.976398 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jszln"] Oct 01 09:39:39 crc kubenswrapper[4787]: I1001 09:39:39.984191 4787 scope.go:117] "RemoveContainer" containerID="a0d13b015eb0406ac50c47a4c50c7ebbeb73445272303b0c75282e8cc39357d4" Oct 01 09:39:39 crc kubenswrapper[4787]: I1001 09:39:39.998983 4787 scope.go:117] "RemoveContainer" containerID="f9ee46727fb7e8eb6dec12809885a91b1ecd1938a84356c816455aa177174f31" Oct 01 09:39:39 crc kubenswrapper[4787]: E1001 09:39:39.999639 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9ee46727fb7e8eb6dec12809885a91b1ecd1938a84356c816455aa177174f31\": container with ID starting with f9ee46727fb7e8eb6dec12809885a91b1ecd1938a84356c816455aa177174f31 not found: ID does not exist" containerID="f9ee46727fb7e8eb6dec12809885a91b1ecd1938a84356c816455aa177174f31" Oct 01 09:39:39 crc kubenswrapper[4787]: I1001 09:39:39.999683 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9ee46727fb7e8eb6dec12809885a91b1ecd1938a84356c816455aa177174f31"} err="failed to get container status \"f9ee46727fb7e8eb6dec12809885a91b1ecd1938a84356c816455aa177174f31\": rpc error: code = NotFound desc = could not find container \"f9ee46727fb7e8eb6dec12809885a91b1ecd1938a84356c816455aa177174f31\": container with ID starting with f9ee46727fb7e8eb6dec12809885a91b1ecd1938a84356c816455aa177174f31 not found: ID does not exist" Oct 01 09:39:39 crc kubenswrapper[4787]: I1001 09:39:39.999712 4787 scope.go:117] "RemoveContainer" containerID="e17a2054049276967a4ee7888f5f986176669408fb2c1fc161f9e0d338de2c68" Oct 01 09:39:40 crc kubenswrapper[4787]: E1001 09:39:40.000538 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e17a2054049276967a4ee7888f5f986176669408fb2c1fc161f9e0d338de2c68\": container with ID starting with e17a2054049276967a4ee7888f5f986176669408fb2c1fc161f9e0d338de2c68 not found: ID does not exist" containerID="e17a2054049276967a4ee7888f5f986176669408fb2c1fc161f9e0d338de2c68" Oct 01 09:39:40 crc kubenswrapper[4787]: I1001 09:39:40.000570 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e17a2054049276967a4ee7888f5f986176669408fb2c1fc161f9e0d338de2c68"} err="failed to get container status \"e17a2054049276967a4ee7888f5f986176669408fb2c1fc161f9e0d338de2c68\": rpc error: code = NotFound desc = could not find container \"e17a2054049276967a4ee7888f5f986176669408fb2c1fc161f9e0d338de2c68\": container with ID starting with e17a2054049276967a4ee7888f5f986176669408fb2c1fc161f9e0d338de2c68 not found: ID does not exist" Oct 01 09:39:40 crc kubenswrapper[4787]: I1001 09:39:40.000594 4787 scope.go:117] "RemoveContainer" containerID="a0d13b015eb0406ac50c47a4c50c7ebbeb73445272303b0c75282e8cc39357d4" Oct 01 09:39:40 crc kubenswrapper[4787]: E1001 09:39:40.000940 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0d13b015eb0406ac50c47a4c50c7ebbeb73445272303b0c75282e8cc39357d4\": container with ID starting with a0d13b015eb0406ac50c47a4c50c7ebbeb73445272303b0c75282e8cc39357d4 not found: ID does not exist" containerID="a0d13b015eb0406ac50c47a4c50c7ebbeb73445272303b0c75282e8cc39357d4" Oct 01 09:39:40 crc kubenswrapper[4787]: I1001 09:39:40.000996 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0d13b015eb0406ac50c47a4c50c7ebbeb73445272303b0c75282e8cc39357d4"} err="failed to get container status \"a0d13b015eb0406ac50c47a4c50c7ebbeb73445272303b0c75282e8cc39357d4\": rpc error: code = NotFound desc = could not find container \"a0d13b015eb0406ac50c47a4c50c7ebbeb73445272303b0c75282e8cc39357d4\": container with ID starting with a0d13b015eb0406ac50c47a4c50c7ebbeb73445272303b0c75282e8cc39357d4 not found: ID does not exist" Oct 01 09:39:40 crc kubenswrapper[4787]: I1001 09:39:40.462978 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tzk9m"] Oct 01 09:39:40 crc kubenswrapper[4787]: I1001 09:39:40.530526 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34c3f769-241a-4df9-9f8d-48308997f85c" path="/var/lib/kubelet/pods/34c3f769-241a-4df9-9f8d-48308997f85c/volumes" Oct 01 09:39:40 crc kubenswrapper[4787]: I1001 09:39:40.940144 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tzk9m" podUID="aa6719df-be6a-4209-9f40-70fbba0465a1" containerName="registry-server" containerID="cri-o://7a07abadeccb594a8a4b687220bc38f001beb285909f3bfa2655ce9f6a7314c8" gracePeriod=2 Oct 01 09:39:41 crc kubenswrapper[4787]: I1001 09:39:41.250733 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:39:41 crc kubenswrapper[4787]: I1001 09:39:41.250801 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:39:41 crc kubenswrapper[4787]: I1001 09:39:41.250847 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" Oct 01 09:39:41 crc kubenswrapper[4787]: I1001 09:39:41.251498 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188"} pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 09:39:41 crc kubenswrapper[4787]: I1001 09:39:41.251555 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" containerID="cri-o://50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188" gracePeriod=600 Oct 01 09:39:41 crc kubenswrapper[4787]: I1001 09:39:41.301507 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tzk9m" Oct 01 09:39:41 crc kubenswrapper[4787]: I1001 09:39:41.362344 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvxmq\" (UniqueName: \"kubernetes.io/projected/aa6719df-be6a-4209-9f40-70fbba0465a1-kube-api-access-jvxmq\") pod \"aa6719df-be6a-4209-9f40-70fbba0465a1\" (UID: \"aa6719df-be6a-4209-9f40-70fbba0465a1\") " Oct 01 09:39:41 crc kubenswrapper[4787]: I1001 09:39:41.362669 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa6719df-be6a-4209-9f40-70fbba0465a1-utilities\") pod \"aa6719df-be6a-4209-9f40-70fbba0465a1\" (UID: \"aa6719df-be6a-4209-9f40-70fbba0465a1\") " Oct 01 09:39:41 crc kubenswrapper[4787]: I1001 09:39:41.362708 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa6719df-be6a-4209-9f40-70fbba0465a1-catalog-content\") pod \"aa6719df-be6a-4209-9f40-70fbba0465a1\" (UID: \"aa6719df-be6a-4209-9f40-70fbba0465a1\") " Oct 01 09:39:41 crc kubenswrapper[4787]: I1001 09:39:41.364230 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa6719df-be6a-4209-9f40-70fbba0465a1-utilities" (OuterVolumeSpecName: "utilities") pod "aa6719df-be6a-4209-9f40-70fbba0465a1" (UID: "aa6719df-be6a-4209-9f40-70fbba0465a1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:39:41 crc kubenswrapper[4787]: I1001 09:39:41.372189 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa6719df-be6a-4209-9f40-70fbba0465a1-kube-api-access-jvxmq" (OuterVolumeSpecName: "kube-api-access-jvxmq") pod "aa6719df-be6a-4209-9f40-70fbba0465a1" (UID: "aa6719df-be6a-4209-9f40-70fbba0465a1"). InnerVolumeSpecName "kube-api-access-jvxmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:39:41 crc kubenswrapper[4787]: I1001 09:39:41.379628 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa6719df-be6a-4209-9f40-70fbba0465a1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aa6719df-be6a-4209-9f40-70fbba0465a1" (UID: "aa6719df-be6a-4209-9f40-70fbba0465a1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:39:41 crc kubenswrapper[4787]: I1001 09:39:41.464139 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvxmq\" (UniqueName: \"kubernetes.io/projected/aa6719df-be6a-4209-9f40-70fbba0465a1-kube-api-access-jvxmq\") on node \"crc\" DevicePath \"\"" Oct 01 09:39:41 crc kubenswrapper[4787]: I1001 09:39:41.464173 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa6719df-be6a-4209-9f40-70fbba0465a1-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:39:41 crc kubenswrapper[4787]: I1001 09:39:41.464187 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa6719df-be6a-4209-9f40-70fbba0465a1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:39:41 crc kubenswrapper[4787]: I1001 09:39:41.954339 4787 generic.go:334] "Generic (PLEG): container finished" podID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerID="50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188" exitCode=0 Oct 01 09:39:41 crc kubenswrapper[4787]: I1001 09:39:41.954766 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" event={"ID":"fa51b883-5f3e-4141-9d4c-37704aac7718","Type":"ContainerDied","Data":"50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188"} Oct 01 09:39:41 crc kubenswrapper[4787]: I1001 09:39:41.954861 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" event={"ID":"fa51b883-5f3e-4141-9d4c-37704aac7718","Type":"ContainerStarted","Data":"bc6fdce923b6e1a3fe85923c737b807469b5f84a3dceb7d7e024941e22a498f2"} Oct 01 09:39:41 crc kubenswrapper[4787]: I1001 09:39:41.960834 4787 generic.go:334] "Generic (PLEG): container finished" podID="aa6719df-be6a-4209-9f40-70fbba0465a1" containerID="7a07abadeccb594a8a4b687220bc38f001beb285909f3bfa2655ce9f6a7314c8" exitCode=0 Oct 01 09:39:41 crc kubenswrapper[4787]: I1001 09:39:41.960884 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tzk9m" event={"ID":"aa6719df-be6a-4209-9f40-70fbba0465a1","Type":"ContainerDied","Data":"7a07abadeccb594a8a4b687220bc38f001beb285909f3bfa2655ce9f6a7314c8"} Oct 01 09:39:41 crc kubenswrapper[4787]: I1001 09:39:41.960915 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tzk9m" event={"ID":"aa6719df-be6a-4209-9f40-70fbba0465a1","Type":"ContainerDied","Data":"b0bf6ed1803f786ade43847557e4c9f42883fee4f25e3f39c57abe637df4e7b1"} Oct 01 09:39:41 crc kubenswrapper[4787]: I1001 09:39:41.961010 4787 scope.go:117] "RemoveContainer" containerID="7a07abadeccb594a8a4b687220bc38f001beb285909f3bfa2655ce9f6a7314c8" Oct 01 09:39:41 crc kubenswrapper[4787]: I1001 09:39:41.961039 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tzk9m" Oct 01 09:39:41 crc kubenswrapper[4787]: I1001 09:39:41.998053 4787 scope.go:117] "RemoveContainer" containerID="2fa9815afd9594d010c212553024c1410df3d8aa855359dcf4bd88ca07d850d3" Oct 01 09:39:42 crc kubenswrapper[4787]: I1001 09:39:42.001339 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tzk9m"] Oct 01 09:39:42 crc kubenswrapper[4787]: I1001 09:39:42.011902 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tzk9m"] Oct 01 09:39:42 crc kubenswrapper[4787]: I1001 09:39:42.027376 4787 scope.go:117] "RemoveContainer" containerID="3b327333988d481b0c9efcac604b3b5b61c97afd9d63c6548eda57f11da0f523" Oct 01 09:39:42 crc kubenswrapper[4787]: I1001 09:39:42.042068 4787 scope.go:117] "RemoveContainer" containerID="7a07abadeccb594a8a4b687220bc38f001beb285909f3bfa2655ce9f6a7314c8" Oct 01 09:39:42 crc kubenswrapper[4787]: E1001 09:39:42.042627 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a07abadeccb594a8a4b687220bc38f001beb285909f3bfa2655ce9f6a7314c8\": container with ID starting with 7a07abadeccb594a8a4b687220bc38f001beb285909f3bfa2655ce9f6a7314c8 not found: ID does not exist" containerID="7a07abadeccb594a8a4b687220bc38f001beb285909f3bfa2655ce9f6a7314c8" Oct 01 09:39:42 crc kubenswrapper[4787]: I1001 09:39:42.042677 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a07abadeccb594a8a4b687220bc38f001beb285909f3bfa2655ce9f6a7314c8"} err="failed to get container status \"7a07abadeccb594a8a4b687220bc38f001beb285909f3bfa2655ce9f6a7314c8\": rpc error: code = NotFound desc = could not find container \"7a07abadeccb594a8a4b687220bc38f001beb285909f3bfa2655ce9f6a7314c8\": container with ID starting with 7a07abadeccb594a8a4b687220bc38f001beb285909f3bfa2655ce9f6a7314c8 not found: ID does not exist" Oct 01 09:39:42 crc kubenswrapper[4787]: I1001 09:39:42.042706 4787 scope.go:117] "RemoveContainer" containerID="2fa9815afd9594d010c212553024c1410df3d8aa855359dcf4bd88ca07d850d3" Oct 01 09:39:42 crc kubenswrapper[4787]: E1001 09:39:42.042956 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fa9815afd9594d010c212553024c1410df3d8aa855359dcf4bd88ca07d850d3\": container with ID starting with 2fa9815afd9594d010c212553024c1410df3d8aa855359dcf4bd88ca07d850d3 not found: ID does not exist" containerID="2fa9815afd9594d010c212553024c1410df3d8aa855359dcf4bd88ca07d850d3" Oct 01 09:39:42 crc kubenswrapper[4787]: I1001 09:39:42.042976 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fa9815afd9594d010c212553024c1410df3d8aa855359dcf4bd88ca07d850d3"} err="failed to get container status \"2fa9815afd9594d010c212553024c1410df3d8aa855359dcf4bd88ca07d850d3\": rpc error: code = NotFound desc = could not find container \"2fa9815afd9594d010c212553024c1410df3d8aa855359dcf4bd88ca07d850d3\": container with ID starting with 2fa9815afd9594d010c212553024c1410df3d8aa855359dcf4bd88ca07d850d3 not found: ID does not exist" Oct 01 09:39:42 crc kubenswrapper[4787]: I1001 09:39:42.042988 4787 scope.go:117] "RemoveContainer" containerID="3b327333988d481b0c9efcac604b3b5b61c97afd9d63c6548eda57f11da0f523" Oct 01 09:39:42 crc kubenswrapper[4787]: E1001 09:39:42.043286 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b327333988d481b0c9efcac604b3b5b61c97afd9d63c6548eda57f11da0f523\": container with ID starting with 3b327333988d481b0c9efcac604b3b5b61c97afd9d63c6548eda57f11da0f523 not found: ID does not exist" containerID="3b327333988d481b0c9efcac604b3b5b61c97afd9d63c6548eda57f11da0f523" Oct 01 09:39:42 crc kubenswrapper[4787]: I1001 09:39:42.043307 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b327333988d481b0c9efcac604b3b5b61c97afd9d63c6548eda57f11da0f523"} err="failed to get container status \"3b327333988d481b0c9efcac604b3b5b61c97afd9d63c6548eda57f11da0f523\": rpc error: code = NotFound desc = could not find container \"3b327333988d481b0c9efcac604b3b5b61c97afd9d63c6548eda57f11da0f523\": container with ID starting with 3b327333988d481b0c9efcac604b3b5b61c97afd9d63c6548eda57f11da0f523 not found: ID does not exist" Oct 01 09:39:42 crc kubenswrapper[4787]: I1001 09:39:42.535302 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa6719df-be6a-4209-9f40-70fbba0465a1" path="/var/lib/kubelet/pods/aa6719df-be6a-4209-9f40-70fbba0465a1/volumes" Oct 01 09:39:58 crc kubenswrapper[4787]: I1001 09:39:58.866953 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" podUID="7ef82a9b-8896-4ff4-ae93-91ea7ed2885e" containerName="oauth-openshift" containerID="cri-o://d68100bc22c5dd781e9f555c26fcc97f3da43c3cc79dc00c997b00804c9dd1e1" gracePeriod=15 Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.061589 4787 generic.go:334] "Generic (PLEG): container finished" podID="7ef82a9b-8896-4ff4-ae93-91ea7ed2885e" containerID="d68100bc22c5dd781e9f555c26fcc97f3da43c3cc79dc00c997b00804c9dd1e1" exitCode=0 Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.061723 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" event={"ID":"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e","Type":"ContainerDied","Data":"d68100bc22c5dd781e9f555c26fcc97f3da43c3cc79dc00c997b00804c9dd1e1"} Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.247733 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.288736 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8"] Oct 01 09:39:59 crc kubenswrapper[4787]: E1001 09:39:59.289044 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ef82a9b-8896-4ff4-ae93-91ea7ed2885e" containerName="oauth-openshift" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.289067 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ef82a9b-8896-4ff4-ae93-91ea7ed2885e" containerName="oauth-openshift" Oct 01 09:39:59 crc kubenswrapper[4787]: E1001 09:39:59.289108 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33bdb919-2c9f-4133-be8b-368dea5b099a" containerName="registry-server" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.289123 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="33bdb919-2c9f-4133-be8b-368dea5b099a" containerName="registry-server" Oct 01 09:39:59 crc kubenswrapper[4787]: E1001 09:39:59.289145 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa6719df-be6a-4209-9f40-70fbba0465a1" containerName="registry-server" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.289155 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa6719df-be6a-4209-9f40-70fbba0465a1" containerName="registry-server" Oct 01 09:39:59 crc kubenswrapper[4787]: E1001 09:39:59.289238 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa6719df-be6a-4209-9f40-70fbba0465a1" containerName="extract-utilities" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.289291 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa6719df-be6a-4209-9f40-70fbba0465a1" containerName="extract-utilities" Oct 01 09:39:59 crc kubenswrapper[4787]: E1001 09:39:59.289307 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34c3f769-241a-4df9-9f8d-48308997f85c" containerName="extract-content" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.289319 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="34c3f769-241a-4df9-9f8d-48308997f85c" containerName="extract-content" Oct 01 09:39:59 crc kubenswrapper[4787]: E1001 09:39:59.289333 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34c3f769-241a-4df9-9f8d-48308997f85c" containerName="extract-utilities" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.289342 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="34c3f769-241a-4df9-9f8d-48308997f85c" containerName="extract-utilities" Oct 01 09:39:59 crc kubenswrapper[4787]: E1001 09:39:59.289354 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa6719df-be6a-4209-9f40-70fbba0465a1" containerName="extract-content" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.289367 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa6719df-be6a-4209-9f40-70fbba0465a1" containerName="extract-content" Oct 01 09:39:59 crc kubenswrapper[4787]: E1001 09:39:59.289385 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33bdb919-2c9f-4133-be8b-368dea5b099a" containerName="extract-utilities" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.289401 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="33bdb919-2c9f-4133-be8b-368dea5b099a" containerName="extract-utilities" Oct 01 09:39:59 crc kubenswrapper[4787]: E1001 09:39:59.289419 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0637215a-7749-4b92-a60e-f24e2d1b1a29" containerName="extract-utilities" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.289429 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="0637215a-7749-4b92-a60e-f24e2d1b1a29" containerName="extract-utilities" Oct 01 09:39:59 crc kubenswrapper[4787]: E1001 09:39:59.289448 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0637215a-7749-4b92-a60e-f24e2d1b1a29" containerName="extract-content" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.289459 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="0637215a-7749-4b92-a60e-f24e2d1b1a29" containerName="extract-content" Oct 01 09:39:59 crc kubenswrapper[4787]: E1001 09:39:59.289478 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ec992e3-443f-4601-9faf-ca18c832a8db" containerName="pruner" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.289489 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ec992e3-443f-4601-9faf-ca18c832a8db" containerName="pruner" Oct 01 09:39:59 crc kubenswrapper[4787]: E1001 09:39:59.289506 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33bdb919-2c9f-4133-be8b-368dea5b099a" containerName="extract-content" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.289517 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="33bdb919-2c9f-4133-be8b-368dea5b099a" containerName="extract-content" Oct 01 09:39:59 crc kubenswrapper[4787]: E1001 09:39:59.289538 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8de4a645-2945-4261-b9be-ee3336e818df" containerName="collect-profiles" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.289548 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="8de4a645-2945-4261-b9be-ee3336e818df" containerName="collect-profiles" Oct 01 09:39:59 crc kubenswrapper[4787]: E1001 09:39:59.289562 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0637215a-7749-4b92-a60e-f24e2d1b1a29" containerName="registry-server" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.289571 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="0637215a-7749-4b92-a60e-f24e2d1b1a29" containerName="registry-server" Oct 01 09:39:59 crc kubenswrapper[4787]: E1001 09:39:59.289582 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa71891b-473a-4cfd-a80a-802a90f14ce4" containerName="pruner" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.289591 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa71891b-473a-4cfd-a80a-802a90f14ce4" containerName="pruner" Oct 01 09:39:59 crc kubenswrapper[4787]: E1001 09:39:59.289602 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34c3f769-241a-4df9-9f8d-48308997f85c" containerName="registry-server" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.289612 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="34c3f769-241a-4df9-9f8d-48308997f85c" containerName="registry-server" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.289747 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa6719df-be6a-4209-9f40-70fbba0465a1" containerName="registry-server" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.289764 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="33bdb919-2c9f-4133-be8b-368dea5b099a" containerName="registry-server" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.289777 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="34c3f769-241a-4df9-9f8d-48308997f85c" containerName="registry-server" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.289791 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="8de4a645-2945-4261-b9be-ee3336e818df" containerName="collect-profiles" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.289802 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ec992e3-443f-4601-9faf-ca18c832a8db" containerName="pruner" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.289820 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa71891b-473a-4cfd-a80a-802a90f14ce4" containerName="pruner" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.289834 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="0637215a-7749-4b92-a60e-f24e2d1b1a29" containerName="registry-server" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.289844 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ef82a9b-8896-4ff4-ae93-91ea7ed2885e" containerName="oauth-openshift" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.290498 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.300783 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-ocp-branding-template\") pod \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.300860 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-router-certs\") pod \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.300913 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-user-idp-0-file-data\") pod \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.301643 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srs6n\" (UniqueName: \"kubernetes.io/projected/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-kube-api-access-srs6n\") pod \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.301708 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-audit-policies\") pod \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.301774 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-user-template-login\") pod \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.301820 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-session\") pod \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.301857 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-user-template-error\") pod \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.301905 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-trusted-ca-bundle\") pod \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.302043 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-serving-cert\") pod \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.302114 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-user-template-provider-selection\") pod \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.302156 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-service-ca\") pod \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.302194 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-cliconfig\") pod \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.302232 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-audit-dir\") pod \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\" (UID: \"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e\") " Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.302585 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "7ef82a9b-8896-4ff4-ae93-91ea7ed2885e" (UID: "7ef82a9b-8896-4ff4-ae93-91ea7ed2885e"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.306414 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "7ef82a9b-8896-4ff4-ae93-91ea7ed2885e" (UID: "7ef82a9b-8896-4ff4-ae93-91ea7ed2885e"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.307034 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "7ef82a9b-8896-4ff4-ae93-91ea7ed2885e" (UID: "7ef82a9b-8896-4ff4-ae93-91ea7ed2885e"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.307560 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "7ef82a9b-8896-4ff4-ae93-91ea7ed2885e" (UID: "7ef82a9b-8896-4ff4-ae93-91ea7ed2885e"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.307938 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "7ef82a9b-8896-4ff4-ae93-91ea7ed2885e" (UID: "7ef82a9b-8896-4ff4-ae93-91ea7ed2885e"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.312694 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "7ef82a9b-8896-4ff4-ae93-91ea7ed2885e" (UID: "7ef82a9b-8896-4ff4-ae93-91ea7ed2885e"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.317479 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "7ef82a9b-8896-4ff4-ae93-91ea7ed2885e" (UID: "7ef82a9b-8896-4ff4-ae93-91ea7ed2885e"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.317791 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "7ef82a9b-8896-4ff4-ae93-91ea7ed2885e" (UID: "7ef82a9b-8896-4ff4-ae93-91ea7ed2885e"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.320871 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "7ef82a9b-8896-4ff4-ae93-91ea7ed2885e" (UID: "7ef82a9b-8896-4ff4-ae93-91ea7ed2885e"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.320945 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "7ef82a9b-8896-4ff4-ae93-91ea7ed2885e" (UID: "7ef82a9b-8896-4ff4-ae93-91ea7ed2885e"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.321302 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-kube-api-access-srs6n" (OuterVolumeSpecName: "kube-api-access-srs6n") pod "7ef82a9b-8896-4ff4-ae93-91ea7ed2885e" (UID: "7ef82a9b-8896-4ff4-ae93-91ea7ed2885e"). InnerVolumeSpecName "kube-api-access-srs6n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.321749 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "7ef82a9b-8896-4ff4-ae93-91ea7ed2885e" (UID: "7ef82a9b-8896-4ff4-ae93-91ea7ed2885e"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.323304 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8"] Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.324725 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "7ef82a9b-8896-4ff4-ae93-91ea7ed2885e" (UID: "7ef82a9b-8896-4ff4-ae93-91ea7ed2885e"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.325510 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "7ef82a9b-8896-4ff4-ae93-91ea7ed2885e" (UID: "7ef82a9b-8896-4ff4-ae93-91ea7ed2885e"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.403993 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-v4-0-config-system-session\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.404127 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-v4-0-config-user-template-error\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.404158 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.404181 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-audit-dir\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.404275 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sl69w\" (UniqueName: \"kubernetes.io/projected/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-kube-api-access-sl69w\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.404317 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.404354 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.404374 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-audit-policies\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.404399 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.404445 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.404512 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-v4-0-config-system-service-ca\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.404582 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-v4-0-config-system-router-certs\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.404654 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-v4-0-config-user-template-login\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.404726 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.404782 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.404801 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.404819 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.404831 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.404845 4787 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.404857 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.404869 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.404882 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.404894 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srs6n\" (UniqueName: \"kubernetes.io/projected/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-kube-api-access-srs6n\") on node \"crc\" DevicePath \"\"" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.404905 4787 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.404929 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.404945 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.404957 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.404968 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.506744 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-v4-0-config-user-template-error\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.506853 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.506902 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-audit-dir\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.506957 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sl69w\" (UniqueName: \"kubernetes.io/projected/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-kube-api-access-sl69w\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.507005 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.507067 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.507127 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-audit-policies\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.507167 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.507198 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.507251 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-v4-0-config-system-service-ca\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.507306 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-v4-0-config-user-template-login\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.507342 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-v4-0-config-system-router-certs\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.507385 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.507423 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-v4-0-config-system-session\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.518481 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.519483 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.507113 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-audit-dir\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.522411 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-v4-0-config-user-template-error\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.522413 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.522888 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-audit-policies\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.523045 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.523992 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-v4-0-config-system-service-ca\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.524411 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.525676 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-v4-0-config-system-session\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.526409 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-v4-0-config-system-router-certs\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.526917 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.531665 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sl69w\" (UniqueName: \"kubernetes.io/projected/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-kube-api-access-sl69w\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.532054 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/d181f4d0-fec2-4eb4-8ab8-60c51e70d705-v4-0-config-user-template-login\") pod \"oauth-openshift-6bbf4c9fdf-hl6g8\" (UID: \"d181f4d0-fec2-4eb4-8ab8-60c51e70d705\") " pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.647026 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:39:59 crc kubenswrapper[4787]: I1001 09:39:59.905499 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8"] Oct 01 09:39:59 crc kubenswrapper[4787]: W1001 09:39:59.912606 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd181f4d0_fec2_4eb4_8ab8_60c51e70d705.slice/crio-ea3f914a9471c3a0f3943d4b87433110e2216758f475ba6d5c0ecc216509e276 WatchSource:0}: Error finding container ea3f914a9471c3a0f3943d4b87433110e2216758f475ba6d5c0ecc216509e276: Status 404 returned error can't find the container with id ea3f914a9471c3a0f3943d4b87433110e2216758f475ba6d5c0ecc216509e276 Oct 01 09:40:00 crc kubenswrapper[4787]: I1001 09:40:00.069707 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" event={"ID":"d181f4d0-fec2-4eb4-8ab8-60c51e70d705","Type":"ContainerStarted","Data":"ea3f914a9471c3a0f3943d4b87433110e2216758f475ba6d5c0ecc216509e276"} Oct 01 09:40:00 crc kubenswrapper[4787]: I1001 09:40:00.071811 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" event={"ID":"7ef82a9b-8896-4ff4-ae93-91ea7ed2885e","Type":"ContainerDied","Data":"192a63e14bec77d4764e90ac496ddeefd96b9dff73c66657597f93f3c37901bb"} Oct 01 09:40:00 crc kubenswrapper[4787]: I1001 09:40:00.071908 4787 scope.go:117] "RemoveContainer" containerID="d68100bc22c5dd781e9f555c26fcc97f3da43c3cc79dc00c997b00804c9dd1e1" Oct 01 09:40:00 crc kubenswrapper[4787]: I1001 09:40:00.072189 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-hwftw" Oct 01 09:40:00 crc kubenswrapper[4787]: I1001 09:40:00.122487 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-hwftw"] Oct 01 09:40:00 crc kubenswrapper[4787]: I1001 09:40:00.127214 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-hwftw"] Oct 01 09:40:00 crc kubenswrapper[4787]: I1001 09:40:00.540776 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ef82a9b-8896-4ff4-ae93-91ea7ed2885e" path="/var/lib/kubelet/pods/7ef82a9b-8896-4ff4-ae93-91ea7ed2885e/volumes" Oct 01 09:40:01 crc kubenswrapper[4787]: I1001 09:40:01.077227 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" event={"ID":"d181f4d0-fec2-4eb4-8ab8-60c51e70d705","Type":"ContainerStarted","Data":"0073695c8440b91723a1434b2811839e09ebfd46fed8853524c3134a5be5080d"} Oct 01 09:40:01 crc kubenswrapper[4787]: I1001 09:40:01.077509 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:40:01 crc kubenswrapper[4787]: I1001 09:40:01.082878 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" Oct 01 09:40:01 crc kubenswrapper[4787]: I1001 09:40:01.099260 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-6bbf4c9fdf-hl6g8" podStartSLOduration=28.099243593 podStartE2EDuration="28.099243593s" podCreationTimestamp="2025-10-01 09:39:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:40:01.095235107 +0000 UTC m=+233.210379274" watchObservedRunningTime="2025-10-01 09:40:01.099243593 +0000 UTC m=+233.214387750" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.004233 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cqr6k"] Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.005255 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-cqr6k" podUID="c4fd105d-9feb-4898-92a1-9c7ae1613202" containerName="registry-server" containerID="cri-o://470dad0110c39dc6d9f70ec7a0c4b95126dcc8304888d81e74292727dfeec681" gracePeriod=30 Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.013727 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xq2kw"] Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.014018 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xq2kw" podUID="045fb03e-9b60-4729-b4cd-73db79bb6294" containerName="registry-server" containerID="cri-o://6e5aa283bce46500c27c7ce2816d6399083309226b0a0c302494b888e82609ba" gracePeriod=30 Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.025987 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cf76m"] Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.026309 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-cf76m" podUID="5d7d263b-bc04-4069-81a5-d5cd1151d805" containerName="marketplace-operator" containerID="cri-o://412ca820d6e3487af9fa549927f8fcb2620256515d9da3df5360333631deaaf2" gracePeriod=30 Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.033778 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z2fsb"] Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.034062 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-z2fsb" podUID="315c2b9e-c2cc-4192-8a74-fe35860ff463" containerName="registry-server" containerID="cri-o://ba3bfd8b91dad1bc5fbd89dc0c8b5bea4a527995ca6d733a5394a85f088cd3e6" gracePeriod=30 Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.039148 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t6rc9"] Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.039361 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-t6rc9" podUID="4914dc93-bedf-4429-a399-9f465e2c0ab9" containerName="registry-server" containerID="cri-o://18558e35e002ca21b2ec847e1bb25fb54f8a3f5605ad161eeb2e007c88f96a22" gracePeriod=30 Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.051750 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6zb6m"] Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.052628 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-6zb6m" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.107752 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6zb6m"] Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.185613 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3390abc0-6474-4a34-8e95-c5923f699e80-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-6zb6m\" (UID: \"3390abc0-6474-4a34-8e95-c5923f699e80\") " pod="openshift-marketplace/marketplace-operator-79b997595-6zb6m" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.186038 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3390abc0-6474-4a34-8e95-c5923f699e80-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-6zb6m\" (UID: \"3390abc0-6474-4a34-8e95-c5923f699e80\") " pod="openshift-marketplace/marketplace-operator-79b997595-6zb6m" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.186194 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qfs9\" (UniqueName: \"kubernetes.io/projected/3390abc0-6474-4a34-8e95-c5923f699e80-kube-api-access-6qfs9\") pod \"marketplace-operator-79b997595-6zb6m\" (UID: \"3390abc0-6474-4a34-8e95-c5923f699e80\") " pod="openshift-marketplace/marketplace-operator-79b997595-6zb6m" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.223199 4787 generic.go:334] "Generic (PLEG): container finished" podID="4914dc93-bedf-4429-a399-9f465e2c0ab9" containerID="18558e35e002ca21b2ec847e1bb25fb54f8a3f5605ad161eeb2e007c88f96a22" exitCode=0 Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.223275 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t6rc9" event={"ID":"4914dc93-bedf-4429-a399-9f465e2c0ab9","Type":"ContainerDied","Data":"18558e35e002ca21b2ec847e1bb25fb54f8a3f5605ad161eeb2e007c88f96a22"} Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.236771 4787 generic.go:334] "Generic (PLEG): container finished" podID="045fb03e-9b60-4729-b4cd-73db79bb6294" containerID="6e5aa283bce46500c27c7ce2816d6399083309226b0a0c302494b888e82609ba" exitCode=0 Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.236877 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xq2kw" event={"ID":"045fb03e-9b60-4729-b4cd-73db79bb6294","Type":"ContainerDied","Data":"6e5aa283bce46500c27c7ce2816d6399083309226b0a0c302494b888e82609ba"} Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.242409 4787 generic.go:334] "Generic (PLEG): container finished" podID="5d7d263b-bc04-4069-81a5-d5cd1151d805" containerID="412ca820d6e3487af9fa549927f8fcb2620256515d9da3df5360333631deaaf2" exitCode=0 Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.242477 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cf76m" event={"ID":"5d7d263b-bc04-4069-81a5-d5cd1151d805","Type":"ContainerDied","Data":"412ca820d6e3487af9fa549927f8fcb2620256515d9da3df5360333631deaaf2"} Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.245776 4787 generic.go:334] "Generic (PLEG): container finished" podID="c4fd105d-9feb-4898-92a1-9c7ae1613202" containerID="470dad0110c39dc6d9f70ec7a0c4b95126dcc8304888d81e74292727dfeec681" exitCode=0 Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.245849 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cqr6k" event={"ID":"c4fd105d-9feb-4898-92a1-9c7ae1613202","Type":"ContainerDied","Data":"470dad0110c39dc6d9f70ec7a0c4b95126dcc8304888d81e74292727dfeec681"} Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.252861 4787 generic.go:334] "Generic (PLEG): container finished" podID="315c2b9e-c2cc-4192-8a74-fe35860ff463" containerID="ba3bfd8b91dad1bc5fbd89dc0c8b5bea4a527995ca6d733a5394a85f088cd3e6" exitCode=0 Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.252889 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z2fsb" event={"ID":"315c2b9e-c2cc-4192-8a74-fe35860ff463","Type":"ContainerDied","Data":"ba3bfd8b91dad1bc5fbd89dc0c8b5bea4a527995ca6d733a5394a85f088cd3e6"} Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.287851 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qfs9\" (UniqueName: \"kubernetes.io/projected/3390abc0-6474-4a34-8e95-c5923f699e80-kube-api-access-6qfs9\") pod \"marketplace-operator-79b997595-6zb6m\" (UID: \"3390abc0-6474-4a34-8e95-c5923f699e80\") " pod="openshift-marketplace/marketplace-operator-79b997595-6zb6m" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.287924 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3390abc0-6474-4a34-8e95-c5923f699e80-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-6zb6m\" (UID: \"3390abc0-6474-4a34-8e95-c5923f699e80\") " pod="openshift-marketplace/marketplace-operator-79b997595-6zb6m" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.287943 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3390abc0-6474-4a34-8e95-c5923f699e80-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-6zb6m\" (UID: \"3390abc0-6474-4a34-8e95-c5923f699e80\") " pod="openshift-marketplace/marketplace-operator-79b997595-6zb6m" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.289606 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3390abc0-6474-4a34-8e95-c5923f699e80-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-6zb6m\" (UID: \"3390abc0-6474-4a34-8e95-c5923f699e80\") " pod="openshift-marketplace/marketplace-operator-79b997595-6zb6m" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.302425 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3390abc0-6474-4a34-8e95-c5923f699e80-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-6zb6m\" (UID: \"3390abc0-6474-4a34-8e95-c5923f699e80\") " pod="openshift-marketplace/marketplace-operator-79b997595-6zb6m" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.306055 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qfs9\" (UniqueName: \"kubernetes.io/projected/3390abc0-6474-4a34-8e95-c5923f699e80-kube-api-access-6qfs9\") pod \"marketplace-operator-79b997595-6zb6m\" (UID: \"3390abc0-6474-4a34-8e95-c5923f699e80\") " pod="openshift-marketplace/marketplace-operator-79b997595-6zb6m" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.368275 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-6zb6m" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.520283 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cqr6k" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.530441 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z2fsb" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.533392 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xq2kw" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.572483 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6zb6m"] Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.579690 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t6rc9" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.591818 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4fd105d-9feb-4898-92a1-9c7ae1613202-catalog-content\") pod \"c4fd105d-9feb-4898-92a1-9c7ae1613202\" (UID: \"c4fd105d-9feb-4898-92a1-9c7ae1613202\") " Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.591927 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/045fb03e-9b60-4729-b4cd-73db79bb6294-utilities\") pod \"045fb03e-9b60-4729-b4cd-73db79bb6294\" (UID: \"045fb03e-9b60-4729-b4cd-73db79bb6294\") " Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.591991 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7fht\" (UniqueName: \"kubernetes.io/projected/c4fd105d-9feb-4898-92a1-9c7ae1613202-kube-api-access-s7fht\") pod \"c4fd105d-9feb-4898-92a1-9c7ae1613202\" (UID: \"c4fd105d-9feb-4898-92a1-9c7ae1613202\") " Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.592090 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m96f7\" (UniqueName: \"kubernetes.io/projected/315c2b9e-c2cc-4192-8a74-fe35860ff463-kube-api-access-m96f7\") pod \"315c2b9e-c2cc-4192-8a74-fe35860ff463\" (UID: \"315c2b9e-c2cc-4192-8a74-fe35860ff463\") " Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.592136 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/315c2b9e-c2cc-4192-8a74-fe35860ff463-catalog-content\") pod \"315c2b9e-c2cc-4192-8a74-fe35860ff463\" (UID: \"315c2b9e-c2cc-4192-8a74-fe35860ff463\") " Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.592169 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4fd105d-9feb-4898-92a1-9c7ae1613202-utilities\") pod \"c4fd105d-9feb-4898-92a1-9c7ae1613202\" (UID: \"c4fd105d-9feb-4898-92a1-9c7ae1613202\") " Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.592211 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/315c2b9e-c2cc-4192-8a74-fe35860ff463-utilities\") pod \"315c2b9e-c2cc-4192-8a74-fe35860ff463\" (UID: \"315c2b9e-c2cc-4192-8a74-fe35860ff463\") " Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.592258 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gpzd6\" (UniqueName: \"kubernetes.io/projected/045fb03e-9b60-4729-b4cd-73db79bb6294-kube-api-access-gpzd6\") pod \"045fb03e-9b60-4729-b4cd-73db79bb6294\" (UID: \"045fb03e-9b60-4729-b4cd-73db79bb6294\") " Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.592308 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/045fb03e-9b60-4729-b4cd-73db79bb6294-catalog-content\") pod \"045fb03e-9b60-4729-b4cd-73db79bb6294\" (UID: \"045fb03e-9b60-4729-b4cd-73db79bb6294\") " Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.594630 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4fd105d-9feb-4898-92a1-9c7ae1613202-utilities" (OuterVolumeSpecName: "utilities") pod "c4fd105d-9feb-4898-92a1-9c7ae1613202" (UID: "c4fd105d-9feb-4898-92a1-9c7ae1613202"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.596398 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/045fb03e-9b60-4729-b4cd-73db79bb6294-utilities" (OuterVolumeSpecName: "utilities") pod "045fb03e-9b60-4729-b4cd-73db79bb6294" (UID: "045fb03e-9b60-4729-b4cd-73db79bb6294"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.601990 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/315c2b9e-c2cc-4192-8a74-fe35860ff463-kube-api-access-m96f7" (OuterVolumeSpecName: "kube-api-access-m96f7") pod "315c2b9e-c2cc-4192-8a74-fe35860ff463" (UID: "315c2b9e-c2cc-4192-8a74-fe35860ff463"). InnerVolumeSpecName "kube-api-access-m96f7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.606250 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/045fb03e-9b60-4729-b4cd-73db79bb6294-kube-api-access-gpzd6" (OuterVolumeSpecName: "kube-api-access-gpzd6") pod "045fb03e-9b60-4729-b4cd-73db79bb6294" (UID: "045fb03e-9b60-4729-b4cd-73db79bb6294"). InnerVolumeSpecName "kube-api-access-gpzd6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.610187 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/315c2b9e-c2cc-4192-8a74-fe35860ff463-utilities" (OuterVolumeSpecName: "utilities") pod "315c2b9e-c2cc-4192-8a74-fe35860ff463" (UID: "315c2b9e-c2cc-4192-8a74-fe35860ff463"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.615814 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4fd105d-9feb-4898-92a1-9c7ae1613202-kube-api-access-s7fht" (OuterVolumeSpecName: "kube-api-access-s7fht") pod "c4fd105d-9feb-4898-92a1-9c7ae1613202" (UID: "c4fd105d-9feb-4898-92a1-9c7ae1613202"). InnerVolumeSpecName "kube-api-access-s7fht". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.627572 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/315c2b9e-c2cc-4192-8a74-fe35860ff463-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "315c2b9e-c2cc-4192-8a74-fe35860ff463" (UID: "315c2b9e-c2cc-4192-8a74-fe35860ff463"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.664957 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4fd105d-9feb-4898-92a1-9c7ae1613202-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c4fd105d-9feb-4898-92a1-9c7ae1613202" (UID: "c4fd105d-9feb-4898-92a1-9c7ae1613202"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.679222 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cf76m" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.692483 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/045fb03e-9b60-4729-b4cd-73db79bb6294-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "045fb03e-9b60-4729-b4cd-73db79bb6294" (UID: "045fb03e-9b60-4729-b4cd-73db79bb6294"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.693708 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4914dc93-bedf-4429-a399-9f465e2c0ab9-utilities\") pod \"4914dc93-bedf-4429-a399-9f465e2c0ab9\" (UID: \"4914dc93-bedf-4429-a399-9f465e2c0ab9\") " Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.694052 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4914dc93-bedf-4429-a399-9f465e2c0ab9-catalog-content\") pod \"4914dc93-bedf-4429-a399-9f465e2c0ab9\" (UID: \"4914dc93-bedf-4429-a399-9f465e2c0ab9\") " Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.694361 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4dxt\" (UniqueName: \"kubernetes.io/projected/4914dc93-bedf-4429-a399-9f465e2c0ab9-kube-api-access-b4dxt\") pod \"4914dc93-bedf-4429-a399-9f465e2c0ab9\" (UID: \"4914dc93-bedf-4429-a399-9f465e2c0ab9\") " Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.694614 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m96f7\" (UniqueName: \"kubernetes.io/projected/315c2b9e-c2cc-4192-8a74-fe35860ff463-kube-api-access-m96f7\") on node \"crc\" DevicePath \"\"" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.694632 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/315c2b9e-c2cc-4192-8a74-fe35860ff463-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.694643 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4fd105d-9feb-4898-92a1-9c7ae1613202-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.694653 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/315c2b9e-c2cc-4192-8a74-fe35860ff463-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.694661 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gpzd6\" (UniqueName: \"kubernetes.io/projected/045fb03e-9b60-4729-b4cd-73db79bb6294-kube-api-access-gpzd6\") on node \"crc\" DevicePath \"\"" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.694670 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/045fb03e-9b60-4729-b4cd-73db79bb6294-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.694678 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4fd105d-9feb-4898-92a1-9c7ae1613202-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.694686 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/045fb03e-9b60-4729-b4cd-73db79bb6294-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.694694 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7fht\" (UniqueName: \"kubernetes.io/projected/c4fd105d-9feb-4898-92a1-9c7ae1613202-kube-api-access-s7fht\") on node \"crc\" DevicePath \"\"" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.695704 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4914dc93-bedf-4429-a399-9f465e2c0ab9-utilities" (OuterVolumeSpecName: "utilities") pod "4914dc93-bedf-4429-a399-9f465e2c0ab9" (UID: "4914dc93-bedf-4429-a399-9f465e2c0ab9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.700534 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4914dc93-bedf-4429-a399-9f465e2c0ab9-kube-api-access-b4dxt" (OuterVolumeSpecName: "kube-api-access-b4dxt") pod "4914dc93-bedf-4429-a399-9f465e2c0ab9" (UID: "4914dc93-bedf-4429-a399-9f465e2c0ab9"). InnerVolumeSpecName "kube-api-access-b4dxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.790300 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4914dc93-bedf-4429-a399-9f465e2c0ab9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4914dc93-bedf-4429-a399-9f465e2c0ab9" (UID: "4914dc93-bedf-4429-a399-9f465e2c0ab9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.795924 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzljp\" (UniqueName: \"kubernetes.io/projected/5d7d263b-bc04-4069-81a5-d5cd1151d805-kube-api-access-gzljp\") pod \"5d7d263b-bc04-4069-81a5-d5cd1151d805\" (UID: \"5d7d263b-bc04-4069-81a5-d5cd1151d805\") " Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.795999 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5d7d263b-bc04-4069-81a5-d5cd1151d805-marketplace-trusted-ca\") pod \"5d7d263b-bc04-4069-81a5-d5cd1151d805\" (UID: \"5d7d263b-bc04-4069-81a5-d5cd1151d805\") " Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.796057 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5d7d263b-bc04-4069-81a5-d5cd1151d805-marketplace-operator-metrics\") pod \"5d7d263b-bc04-4069-81a5-d5cd1151d805\" (UID: \"5d7d263b-bc04-4069-81a5-d5cd1151d805\") " Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.796301 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4914dc93-bedf-4429-a399-9f465e2c0ab9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.796320 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4dxt\" (UniqueName: \"kubernetes.io/projected/4914dc93-bedf-4429-a399-9f465e2c0ab9-kube-api-access-b4dxt\") on node \"crc\" DevicePath \"\"" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.796331 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4914dc93-bedf-4429-a399-9f465e2c0ab9-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.796842 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d7d263b-bc04-4069-81a5-d5cd1151d805-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "5d7d263b-bc04-4069-81a5-d5cd1151d805" (UID: "5d7d263b-bc04-4069-81a5-d5cd1151d805"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.799997 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d7d263b-bc04-4069-81a5-d5cd1151d805-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "5d7d263b-bc04-4069-81a5-d5cd1151d805" (UID: "5d7d263b-bc04-4069-81a5-d5cd1151d805"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.801162 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d7d263b-bc04-4069-81a5-d5cd1151d805-kube-api-access-gzljp" (OuterVolumeSpecName: "kube-api-access-gzljp") pod "5d7d263b-bc04-4069-81a5-d5cd1151d805" (UID: "5d7d263b-bc04-4069-81a5-d5cd1151d805"). InnerVolumeSpecName "kube-api-access-gzljp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.897946 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzljp\" (UniqueName: \"kubernetes.io/projected/5d7d263b-bc04-4069-81a5-d5cd1151d805-kube-api-access-gzljp\") on node \"crc\" DevicePath \"\"" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.897988 4787 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5d7d263b-bc04-4069-81a5-d5cd1151d805-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:40:28 crc kubenswrapper[4787]: I1001 09:40:28.897998 4787 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5d7d263b-bc04-4069-81a5-d5cd1151d805-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 01 09:40:29 crc kubenswrapper[4787]: I1001 09:40:29.260765 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cqr6k" event={"ID":"c4fd105d-9feb-4898-92a1-9c7ae1613202","Type":"ContainerDied","Data":"77064cc1d6b633d8c53eeaf5d1957a67a410973e670c0c8c0115072b30a81249"} Oct 01 09:40:29 crc kubenswrapper[4787]: I1001 09:40:29.261033 4787 scope.go:117] "RemoveContainer" containerID="470dad0110c39dc6d9f70ec7a0c4b95126dcc8304888d81e74292727dfeec681" Oct 01 09:40:29 crc kubenswrapper[4787]: I1001 09:40:29.260805 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cqr6k" Oct 01 09:40:29 crc kubenswrapper[4787]: I1001 09:40:29.264832 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t6rc9" event={"ID":"4914dc93-bedf-4429-a399-9f465e2c0ab9","Type":"ContainerDied","Data":"72fef4d89d69491f191f474d85600c9ef288790c967779d310570c0bed16819d"} Oct 01 09:40:29 crc kubenswrapper[4787]: I1001 09:40:29.264915 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t6rc9" Oct 01 09:40:29 crc kubenswrapper[4787]: I1001 09:40:29.266581 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-6zb6m" event={"ID":"3390abc0-6474-4a34-8e95-c5923f699e80","Type":"ContainerStarted","Data":"2454a6960058c8721ad3422ad1259dd33c245499fb429fe6b087df8b8e7eae24"} Oct 01 09:40:29 crc kubenswrapper[4787]: I1001 09:40:29.266642 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-6zb6m" event={"ID":"3390abc0-6474-4a34-8e95-c5923f699e80","Type":"ContainerStarted","Data":"1006ca66ee765dc94ccdde8a046ec44994ff08f243c0c7652818ad764efe24fc"} Oct 01 09:40:29 crc kubenswrapper[4787]: I1001 09:40:29.267540 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-6zb6m" Oct 01 09:40:29 crc kubenswrapper[4787]: I1001 09:40:29.271123 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xq2kw" event={"ID":"045fb03e-9b60-4729-b4cd-73db79bb6294","Type":"ContainerDied","Data":"e62b65509b8f464082ef0e2f8df708bb9c15c57f13257e30d8bc208e746b154a"} Oct 01 09:40:29 crc kubenswrapper[4787]: I1001 09:40:29.271222 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xq2kw" Oct 01 09:40:29 crc kubenswrapper[4787]: I1001 09:40:29.276229 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-6zb6m" Oct 01 09:40:29 crc kubenswrapper[4787]: I1001 09:40:29.280503 4787 scope.go:117] "RemoveContainer" containerID="939dca2b60d466af76455a0b625d2177e4f2871161a5176c3631d1824a7d7d79" Oct 01 09:40:29 crc kubenswrapper[4787]: I1001 09:40:29.282212 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z2fsb" Oct 01 09:40:29 crc kubenswrapper[4787]: I1001 09:40:29.282172 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z2fsb" event={"ID":"315c2b9e-c2cc-4192-8a74-fe35860ff463","Type":"ContainerDied","Data":"382db25a103803d1ab0d9d3dea0c7e40a10df32e543b3365854a351972ffac73"} Oct 01 09:40:29 crc kubenswrapper[4787]: I1001 09:40:29.285398 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cf76m" event={"ID":"5d7d263b-bc04-4069-81a5-d5cd1151d805","Type":"ContainerDied","Data":"ad1ffd81aa54a22f70acb1266afe5bb2e53d7b7561d798f32bdef1c2f7c3fd5c"} Oct 01 09:40:29 crc kubenswrapper[4787]: I1001 09:40:29.285441 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cf76m" Oct 01 09:40:29 crc kubenswrapper[4787]: I1001 09:40:29.290943 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-6zb6m" podStartSLOduration=1.290925211 podStartE2EDuration="1.290925211s" podCreationTimestamp="2025-10-01 09:40:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:40:29.290397778 +0000 UTC m=+261.405541935" watchObservedRunningTime="2025-10-01 09:40:29.290925211 +0000 UTC m=+261.406069368" Oct 01 09:40:29 crc kubenswrapper[4787]: I1001 09:40:29.348324 4787 scope.go:117] "RemoveContainer" containerID="ba462dbf29972e8af033234af97d7fca424f7e6f2f9296e4a1fe27e84e570c98" Oct 01 09:40:29 crc kubenswrapper[4787]: I1001 09:40:29.359812 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cqr6k"] Oct 01 09:40:29 crc kubenswrapper[4787]: I1001 09:40:29.371687 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-cqr6k"] Oct 01 09:40:29 crc kubenswrapper[4787]: I1001 09:40:29.373495 4787 scope.go:117] "RemoveContainer" containerID="18558e35e002ca21b2ec847e1bb25fb54f8a3f5605ad161eeb2e007c88f96a22" Oct 01 09:40:29 crc kubenswrapper[4787]: I1001 09:40:29.378204 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xq2kw"] Oct 01 09:40:29 crc kubenswrapper[4787]: I1001 09:40:29.382626 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xq2kw"] Oct 01 09:40:29 crc kubenswrapper[4787]: I1001 09:40:29.384486 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t6rc9"] Oct 01 09:40:29 crc kubenswrapper[4787]: I1001 09:40:29.386831 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-t6rc9"] Oct 01 09:40:29 crc kubenswrapper[4787]: I1001 09:40:29.393574 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cf76m"] Oct 01 09:40:29 crc kubenswrapper[4787]: I1001 09:40:29.396616 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cf76m"] Oct 01 09:40:29 crc kubenswrapper[4787]: I1001 09:40:29.398641 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z2fsb"] Oct 01 09:40:29 crc kubenswrapper[4787]: I1001 09:40:29.400949 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-z2fsb"] Oct 01 09:40:29 crc kubenswrapper[4787]: I1001 09:40:29.404601 4787 scope.go:117] "RemoveContainer" containerID="165cb69676439a25be6b8ad9e9a51e2d18e2a76c427cb8b2f3e278f638b28260" Oct 01 09:40:29 crc kubenswrapper[4787]: I1001 09:40:29.426478 4787 scope.go:117] "RemoveContainer" containerID="5671c7654e169dd5e940a2d4607a2cc33bc39e383de818bb73a97c4bc3fece1b" Oct 01 09:40:29 crc kubenswrapper[4787]: I1001 09:40:29.438230 4787 scope.go:117] "RemoveContainer" containerID="6e5aa283bce46500c27c7ce2816d6399083309226b0a0c302494b888e82609ba" Oct 01 09:40:29 crc kubenswrapper[4787]: I1001 09:40:29.453144 4787 scope.go:117] "RemoveContainer" containerID="970bc197a3226fc571cfc9c8e6c8725e78105d5112b1bcb73de8f125d6a62801" Oct 01 09:40:29 crc kubenswrapper[4787]: I1001 09:40:29.472579 4787 scope.go:117] "RemoveContainer" containerID="e440081149f9e943d08a9704521ed7cd0ad44c46e0a65bed4170f50a62fd7e78" Oct 01 09:40:29 crc kubenswrapper[4787]: I1001 09:40:29.485395 4787 scope.go:117] "RemoveContainer" containerID="ba3bfd8b91dad1bc5fbd89dc0c8b5bea4a527995ca6d733a5394a85f088cd3e6" Oct 01 09:40:29 crc kubenswrapper[4787]: I1001 09:40:29.497749 4787 scope.go:117] "RemoveContainer" containerID="e69780c2e3002e5e68b1853ff6e3ce99d978d5cbaf6c1a5da0bd01924665f584" Oct 01 09:40:29 crc kubenswrapper[4787]: I1001 09:40:29.510650 4787 scope.go:117] "RemoveContainer" containerID="107f12d0b4ce8bec9f0bbc58687db3f5e3662af34817e98d5b27c328da014fa2" Oct 01 09:40:29 crc kubenswrapper[4787]: I1001 09:40:29.523775 4787 scope.go:117] "RemoveContainer" containerID="412ca820d6e3487af9fa549927f8fcb2620256515d9da3df5360333631deaaf2" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.220170 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-swbhm"] Oct 01 09:40:30 crc kubenswrapper[4787]: E1001 09:40:30.220372 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4fd105d-9feb-4898-92a1-9c7ae1613202" containerName="extract-content" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.220385 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4fd105d-9feb-4898-92a1-9c7ae1613202" containerName="extract-content" Oct 01 09:40:30 crc kubenswrapper[4787]: E1001 09:40:30.220393 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="315c2b9e-c2cc-4192-8a74-fe35860ff463" containerName="registry-server" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.220398 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="315c2b9e-c2cc-4192-8a74-fe35860ff463" containerName="registry-server" Oct 01 09:40:30 crc kubenswrapper[4787]: E1001 09:40:30.220408 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="045fb03e-9b60-4729-b4cd-73db79bb6294" containerName="extract-content" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.220413 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="045fb03e-9b60-4729-b4cd-73db79bb6294" containerName="extract-content" Oct 01 09:40:30 crc kubenswrapper[4787]: E1001 09:40:30.220423 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="315c2b9e-c2cc-4192-8a74-fe35860ff463" containerName="extract-utilities" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.220429 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="315c2b9e-c2cc-4192-8a74-fe35860ff463" containerName="extract-utilities" Oct 01 09:40:30 crc kubenswrapper[4787]: E1001 09:40:30.220439 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4914dc93-bedf-4429-a399-9f465e2c0ab9" containerName="extract-content" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.220445 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="4914dc93-bedf-4429-a399-9f465e2c0ab9" containerName="extract-content" Oct 01 09:40:30 crc kubenswrapper[4787]: E1001 09:40:30.220454 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4fd105d-9feb-4898-92a1-9c7ae1613202" containerName="registry-server" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.220460 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4fd105d-9feb-4898-92a1-9c7ae1613202" containerName="registry-server" Oct 01 09:40:30 crc kubenswrapper[4787]: E1001 09:40:30.220473 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="315c2b9e-c2cc-4192-8a74-fe35860ff463" containerName="extract-content" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.220478 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="315c2b9e-c2cc-4192-8a74-fe35860ff463" containerName="extract-content" Oct 01 09:40:30 crc kubenswrapper[4787]: E1001 09:40:30.220486 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4914dc93-bedf-4429-a399-9f465e2c0ab9" containerName="extract-utilities" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.220491 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="4914dc93-bedf-4429-a399-9f465e2c0ab9" containerName="extract-utilities" Oct 01 09:40:30 crc kubenswrapper[4787]: E1001 09:40:30.220499 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4914dc93-bedf-4429-a399-9f465e2c0ab9" containerName="registry-server" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.220505 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="4914dc93-bedf-4429-a399-9f465e2c0ab9" containerName="registry-server" Oct 01 09:40:30 crc kubenswrapper[4787]: E1001 09:40:30.220514 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d7d263b-bc04-4069-81a5-d5cd1151d805" containerName="marketplace-operator" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.220520 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d7d263b-bc04-4069-81a5-d5cd1151d805" containerName="marketplace-operator" Oct 01 09:40:30 crc kubenswrapper[4787]: E1001 09:40:30.220527 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="045fb03e-9b60-4729-b4cd-73db79bb6294" containerName="registry-server" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.220532 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="045fb03e-9b60-4729-b4cd-73db79bb6294" containerName="registry-server" Oct 01 09:40:30 crc kubenswrapper[4787]: E1001 09:40:30.220540 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="045fb03e-9b60-4729-b4cd-73db79bb6294" containerName="extract-utilities" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.220546 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="045fb03e-9b60-4729-b4cd-73db79bb6294" containerName="extract-utilities" Oct 01 09:40:30 crc kubenswrapper[4787]: E1001 09:40:30.220557 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4fd105d-9feb-4898-92a1-9c7ae1613202" containerName="extract-utilities" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.220563 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4fd105d-9feb-4898-92a1-9c7ae1613202" containerName="extract-utilities" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.220642 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="045fb03e-9b60-4729-b4cd-73db79bb6294" containerName="registry-server" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.220650 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d7d263b-bc04-4069-81a5-d5cd1151d805" containerName="marketplace-operator" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.220661 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="315c2b9e-c2cc-4192-8a74-fe35860ff463" containerName="registry-server" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.220670 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="4914dc93-bedf-4429-a399-9f465e2c0ab9" containerName="registry-server" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.220679 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4fd105d-9feb-4898-92a1-9c7ae1613202" containerName="registry-server" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.221372 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-swbhm" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.224273 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.227952 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-swbhm"] Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.313508 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5cwb\" (UniqueName: \"kubernetes.io/projected/593f4070-9419-4e2f-9da0-741e86877588-kube-api-access-l5cwb\") pod \"redhat-marketplace-swbhm\" (UID: \"593f4070-9419-4e2f-9da0-741e86877588\") " pod="openshift-marketplace/redhat-marketplace-swbhm" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.314564 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/593f4070-9419-4e2f-9da0-741e86877588-utilities\") pod \"redhat-marketplace-swbhm\" (UID: \"593f4070-9419-4e2f-9da0-741e86877588\") " pod="openshift-marketplace/redhat-marketplace-swbhm" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.314688 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/593f4070-9419-4e2f-9da0-741e86877588-catalog-content\") pod \"redhat-marketplace-swbhm\" (UID: \"593f4070-9419-4e2f-9da0-741e86877588\") " pod="openshift-marketplace/redhat-marketplace-swbhm" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.417170 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5cwb\" (UniqueName: \"kubernetes.io/projected/593f4070-9419-4e2f-9da0-741e86877588-kube-api-access-l5cwb\") pod \"redhat-marketplace-swbhm\" (UID: \"593f4070-9419-4e2f-9da0-741e86877588\") " pod="openshift-marketplace/redhat-marketplace-swbhm" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.417230 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/593f4070-9419-4e2f-9da0-741e86877588-utilities\") pod \"redhat-marketplace-swbhm\" (UID: \"593f4070-9419-4e2f-9da0-741e86877588\") " pod="openshift-marketplace/redhat-marketplace-swbhm" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.417751 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/593f4070-9419-4e2f-9da0-741e86877588-catalog-content\") pod \"redhat-marketplace-swbhm\" (UID: \"593f4070-9419-4e2f-9da0-741e86877588\") " pod="openshift-marketplace/redhat-marketplace-swbhm" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.418347 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/593f4070-9419-4e2f-9da0-741e86877588-catalog-content\") pod \"redhat-marketplace-swbhm\" (UID: \"593f4070-9419-4e2f-9da0-741e86877588\") " pod="openshift-marketplace/redhat-marketplace-swbhm" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.418647 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/593f4070-9419-4e2f-9da0-741e86877588-utilities\") pod \"redhat-marketplace-swbhm\" (UID: \"593f4070-9419-4e2f-9da0-741e86877588\") " pod="openshift-marketplace/redhat-marketplace-swbhm" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.426468 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pwj4z"] Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.428050 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pwj4z" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.433181 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.439027 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pwj4z"] Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.442777 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5cwb\" (UniqueName: \"kubernetes.io/projected/593f4070-9419-4e2f-9da0-741e86877588-kube-api-access-l5cwb\") pod \"redhat-marketplace-swbhm\" (UID: \"593f4070-9419-4e2f-9da0-741e86877588\") " pod="openshift-marketplace/redhat-marketplace-swbhm" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.519566 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klk8h\" (UniqueName: \"kubernetes.io/projected/c3868e91-d922-4768-af6a-123795df5a28-kube-api-access-klk8h\") pod \"redhat-operators-pwj4z\" (UID: \"c3868e91-d922-4768-af6a-123795df5a28\") " pod="openshift-marketplace/redhat-operators-pwj4z" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.519644 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3868e91-d922-4768-af6a-123795df5a28-utilities\") pod \"redhat-operators-pwj4z\" (UID: \"c3868e91-d922-4768-af6a-123795df5a28\") " pod="openshift-marketplace/redhat-operators-pwj4z" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.519672 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3868e91-d922-4768-af6a-123795df5a28-catalog-content\") pod \"redhat-operators-pwj4z\" (UID: \"c3868e91-d922-4768-af6a-123795df5a28\") " pod="openshift-marketplace/redhat-operators-pwj4z" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.533054 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="045fb03e-9b60-4729-b4cd-73db79bb6294" path="/var/lib/kubelet/pods/045fb03e-9b60-4729-b4cd-73db79bb6294/volumes" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.533660 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="315c2b9e-c2cc-4192-8a74-fe35860ff463" path="/var/lib/kubelet/pods/315c2b9e-c2cc-4192-8a74-fe35860ff463/volumes" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.534256 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4914dc93-bedf-4429-a399-9f465e2c0ab9" path="/var/lib/kubelet/pods/4914dc93-bedf-4429-a399-9f465e2c0ab9/volumes" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.534838 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d7d263b-bc04-4069-81a5-d5cd1151d805" path="/var/lib/kubelet/pods/5d7d263b-bc04-4069-81a5-d5cd1151d805/volumes" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.535870 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4fd105d-9feb-4898-92a1-9c7ae1613202" path="/var/lib/kubelet/pods/c4fd105d-9feb-4898-92a1-9c7ae1613202/volumes" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.544764 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-swbhm" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.620709 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klk8h\" (UniqueName: \"kubernetes.io/projected/c3868e91-d922-4768-af6a-123795df5a28-kube-api-access-klk8h\") pod \"redhat-operators-pwj4z\" (UID: \"c3868e91-d922-4768-af6a-123795df5a28\") " pod="openshift-marketplace/redhat-operators-pwj4z" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.620809 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3868e91-d922-4768-af6a-123795df5a28-utilities\") pod \"redhat-operators-pwj4z\" (UID: \"c3868e91-d922-4768-af6a-123795df5a28\") " pod="openshift-marketplace/redhat-operators-pwj4z" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.620832 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3868e91-d922-4768-af6a-123795df5a28-catalog-content\") pod \"redhat-operators-pwj4z\" (UID: \"c3868e91-d922-4768-af6a-123795df5a28\") " pod="openshift-marketplace/redhat-operators-pwj4z" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.621599 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3868e91-d922-4768-af6a-123795df5a28-catalog-content\") pod \"redhat-operators-pwj4z\" (UID: \"c3868e91-d922-4768-af6a-123795df5a28\") " pod="openshift-marketplace/redhat-operators-pwj4z" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.622335 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3868e91-d922-4768-af6a-123795df5a28-utilities\") pod \"redhat-operators-pwj4z\" (UID: \"c3868e91-d922-4768-af6a-123795df5a28\") " pod="openshift-marketplace/redhat-operators-pwj4z" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.646217 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klk8h\" (UniqueName: \"kubernetes.io/projected/c3868e91-d922-4768-af6a-123795df5a28-kube-api-access-klk8h\") pod \"redhat-operators-pwj4z\" (UID: \"c3868e91-d922-4768-af6a-123795df5a28\") " pod="openshift-marketplace/redhat-operators-pwj4z" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.760910 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pwj4z" Oct 01 09:40:30 crc kubenswrapper[4787]: I1001 09:40:30.968838 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-swbhm"] Oct 01 09:40:30 crc kubenswrapper[4787]: W1001 09:40:30.973441 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod593f4070_9419_4e2f_9da0_741e86877588.slice/crio-3e9c322d0cf5aa3fc11ea1147a090184e05fda20a39a7363984cbac36bfdb482 WatchSource:0}: Error finding container 3e9c322d0cf5aa3fc11ea1147a090184e05fda20a39a7363984cbac36bfdb482: Status 404 returned error can't find the container with id 3e9c322d0cf5aa3fc11ea1147a090184e05fda20a39a7363984cbac36bfdb482 Oct 01 09:40:31 crc kubenswrapper[4787]: I1001 09:40:31.132394 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pwj4z"] Oct 01 09:40:31 crc kubenswrapper[4787]: W1001 09:40:31.174022 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc3868e91_d922_4768_af6a_123795df5a28.slice/crio-a8d82567cba5ac430f8c193324404e4e385d5e4078b9e40dadbca8db3c6b2776 WatchSource:0}: Error finding container a8d82567cba5ac430f8c193324404e4e385d5e4078b9e40dadbca8db3c6b2776: Status 404 returned error can't find the container with id a8d82567cba5ac430f8c193324404e4e385d5e4078b9e40dadbca8db3c6b2776 Oct 01 09:40:31 crc kubenswrapper[4787]: I1001 09:40:31.311203 4787 generic.go:334] "Generic (PLEG): container finished" podID="c3868e91-d922-4768-af6a-123795df5a28" containerID="37c798c8be3cda42264255cd41cd0f77191d47faf21c7f84f3c1da1f27db6416" exitCode=0 Oct 01 09:40:31 crc kubenswrapper[4787]: I1001 09:40:31.311263 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pwj4z" event={"ID":"c3868e91-d922-4768-af6a-123795df5a28","Type":"ContainerDied","Data":"37c798c8be3cda42264255cd41cd0f77191d47faf21c7f84f3c1da1f27db6416"} Oct 01 09:40:31 crc kubenswrapper[4787]: I1001 09:40:31.311315 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pwj4z" event={"ID":"c3868e91-d922-4768-af6a-123795df5a28","Type":"ContainerStarted","Data":"a8d82567cba5ac430f8c193324404e4e385d5e4078b9e40dadbca8db3c6b2776"} Oct 01 09:40:31 crc kubenswrapper[4787]: I1001 09:40:31.312489 4787 generic.go:334] "Generic (PLEG): container finished" podID="593f4070-9419-4e2f-9da0-741e86877588" containerID="5e6daceb17d06063ab149350954ba6bb483a715d05977233c9ef9b7816af7677" exitCode=0 Oct 01 09:40:31 crc kubenswrapper[4787]: I1001 09:40:31.312586 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-swbhm" event={"ID":"593f4070-9419-4e2f-9da0-741e86877588","Type":"ContainerDied","Data":"5e6daceb17d06063ab149350954ba6bb483a715d05977233c9ef9b7816af7677"} Oct 01 09:40:31 crc kubenswrapper[4787]: I1001 09:40:31.312630 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-swbhm" event={"ID":"593f4070-9419-4e2f-9da0-741e86877588","Type":"ContainerStarted","Data":"3e9c322d0cf5aa3fc11ea1147a090184e05fda20a39a7363984cbac36bfdb482"} Oct 01 09:40:32 crc kubenswrapper[4787]: I1001 09:40:32.319780 4787 generic.go:334] "Generic (PLEG): container finished" podID="593f4070-9419-4e2f-9da0-741e86877588" containerID="2f2f124313ee5405395305c47b0c27c38b2ac55a0eca34a9e5672914447a948d" exitCode=0 Oct 01 09:40:32 crc kubenswrapper[4787]: I1001 09:40:32.319975 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-swbhm" event={"ID":"593f4070-9419-4e2f-9da0-741e86877588","Type":"ContainerDied","Data":"2f2f124313ee5405395305c47b0c27c38b2ac55a0eca34a9e5672914447a948d"} Oct 01 09:40:32 crc kubenswrapper[4787]: I1001 09:40:32.322011 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pwj4z" event={"ID":"c3868e91-d922-4768-af6a-123795df5a28","Type":"ContainerStarted","Data":"8d0f3728cdd32b7b8f824f48ab5b5ce5f78722e04995e424875a96a8726f3e03"} Oct 01 09:40:32 crc kubenswrapper[4787]: I1001 09:40:32.617274 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8ksjp"] Oct 01 09:40:32 crc kubenswrapper[4787]: I1001 09:40:32.618326 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8ksjp" Oct 01 09:40:32 crc kubenswrapper[4787]: I1001 09:40:32.622705 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 01 09:40:32 crc kubenswrapper[4787]: I1001 09:40:32.637615 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8ksjp"] Oct 01 09:40:32 crc kubenswrapper[4787]: I1001 09:40:32.751142 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jzd8\" (UniqueName: \"kubernetes.io/projected/09d257f5-ca13-42c9-92f8-920f99ebadc1-kube-api-access-5jzd8\") pod \"certified-operators-8ksjp\" (UID: \"09d257f5-ca13-42c9-92f8-920f99ebadc1\") " pod="openshift-marketplace/certified-operators-8ksjp" Oct 01 09:40:32 crc kubenswrapper[4787]: I1001 09:40:32.751210 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09d257f5-ca13-42c9-92f8-920f99ebadc1-utilities\") pod \"certified-operators-8ksjp\" (UID: \"09d257f5-ca13-42c9-92f8-920f99ebadc1\") " pod="openshift-marketplace/certified-operators-8ksjp" Oct 01 09:40:32 crc kubenswrapper[4787]: I1001 09:40:32.751251 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09d257f5-ca13-42c9-92f8-920f99ebadc1-catalog-content\") pod \"certified-operators-8ksjp\" (UID: \"09d257f5-ca13-42c9-92f8-920f99ebadc1\") " pod="openshift-marketplace/certified-operators-8ksjp" Oct 01 09:40:32 crc kubenswrapper[4787]: I1001 09:40:32.823533 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4m4pm"] Oct 01 09:40:32 crc kubenswrapper[4787]: I1001 09:40:32.827089 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4m4pm" Oct 01 09:40:32 crc kubenswrapper[4787]: I1001 09:40:32.831027 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 01 09:40:32 crc kubenswrapper[4787]: I1001 09:40:32.833727 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4m4pm"] Oct 01 09:40:32 crc kubenswrapper[4787]: I1001 09:40:32.852168 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jzd8\" (UniqueName: \"kubernetes.io/projected/09d257f5-ca13-42c9-92f8-920f99ebadc1-kube-api-access-5jzd8\") pod \"certified-operators-8ksjp\" (UID: \"09d257f5-ca13-42c9-92f8-920f99ebadc1\") " pod="openshift-marketplace/certified-operators-8ksjp" Oct 01 09:40:32 crc kubenswrapper[4787]: I1001 09:40:32.852224 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09d257f5-ca13-42c9-92f8-920f99ebadc1-utilities\") pod \"certified-operators-8ksjp\" (UID: \"09d257f5-ca13-42c9-92f8-920f99ebadc1\") " pod="openshift-marketplace/certified-operators-8ksjp" Oct 01 09:40:32 crc kubenswrapper[4787]: I1001 09:40:32.852264 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09d257f5-ca13-42c9-92f8-920f99ebadc1-catalog-content\") pod \"certified-operators-8ksjp\" (UID: \"09d257f5-ca13-42c9-92f8-920f99ebadc1\") " pod="openshift-marketplace/certified-operators-8ksjp" Oct 01 09:40:32 crc kubenswrapper[4787]: I1001 09:40:32.852730 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09d257f5-ca13-42c9-92f8-920f99ebadc1-catalog-content\") pod \"certified-operators-8ksjp\" (UID: \"09d257f5-ca13-42c9-92f8-920f99ebadc1\") " pod="openshift-marketplace/certified-operators-8ksjp" Oct 01 09:40:32 crc kubenswrapper[4787]: I1001 09:40:32.854270 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09d257f5-ca13-42c9-92f8-920f99ebadc1-utilities\") pod \"certified-operators-8ksjp\" (UID: \"09d257f5-ca13-42c9-92f8-920f99ebadc1\") " pod="openshift-marketplace/certified-operators-8ksjp" Oct 01 09:40:32 crc kubenswrapper[4787]: I1001 09:40:32.872551 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jzd8\" (UniqueName: \"kubernetes.io/projected/09d257f5-ca13-42c9-92f8-920f99ebadc1-kube-api-access-5jzd8\") pod \"certified-operators-8ksjp\" (UID: \"09d257f5-ca13-42c9-92f8-920f99ebadc1\") " pod="openshift-marketplace/certified-operators-8ksjp" Oct 01 09:40:32 crc kubenswrapper[4787]: I1001 09:40:32.953158 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skt5t\" (UniqueName: \"kubernetes.io/projected/6e4be47d-87d5-469d-9ad2-b12efb75e1aa-kube-api-access-skt5t\") pod \"community-operators-4m4pm\" (UID: \"6e4be47d-87d5-469d-9ad2-b12efb75e1aa\") " pod="openshift-marketplace/community-operators-4m4pm" Oct 01 09:40:32 crc kubenswrapper[4787]: I1001 09:40:32.953246 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e4be47d-87d5-469d-9ad2-b12efb75e1aa-utilities\") pod \"community-operators-4m4pm\" (UID: \"6e4be47d-87d5-469d-9ad2-b12efb75e1aa\") " pod="openshift-marketplace/community-operators-4m4pm" Oct 01 09:40:32 crc kubenswrapper[4787]: I1001 09:40:32.953280 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e4be47d-87d5-469d-9ad2-b12efb75e1aa-catalog-content\") pod \"community-operators-4m4pm\" (UID: \"6e4be47d-87d5-469d-9ad2-b12efb75e1aa\") " pod="openshift-marketplace/community-operators-4m4pm" Oct 01 09:40:32 crc kubenswrapper[4787]: I1001 09:40:32.995316 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8ksjp" Oct 01 09:40:33 crc kubenswrapper[4787]: I1001 09:40:33.054497 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skt5t\" (UniqueName: \"kubernetes.io/projected/6e4be47d-87d5-469d-9ad2-b12efb75e1aa-kube-api-access-skt5t\") pod \"community-operators-4m4pm\" (UID: \"6e4be47d-87d5-469d-9ad2-b12efb75e1aa\") " pod="openshift-marketplace/community-operators-4m4pm" Oct 01 09:40:33 crc kubenswrapper[4787]: I1001 09:40:33.054581 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e4be47d-87d5-469d-9ad2-b12efb75e1aa-utilities\") pod \"community-operators-4m4pm\" (UID: \"6e4be47d-87d5-469d-9ad2-b12efb75e1aa\") " pod="openshift-marketplace/community-operators-4m4pm" Oct 01 09:40:33 crc kubenswrapper[4787]: I1001 09:40:33.054622 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e4be47d-87d5-469d-9ad2-b12efb75e1aa-catalog-content\") pod \"community-operators-4m4pm\" (UID: \"6e4be47d-87d5-469d-9ad2-b12efb75e1aa\") " pod="openshift-marketplace/community-operators-4m4pm" Oct 01 09:40:33 crc kubenswrapper[4787]: I1001 09:40:33.055696 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e4be47d-87d5-469d-9ad2-b12efb75e1aa-catalog-content\") pod \"community-operators-4m4pm\" (UID: \"6e4be47d-87d5-469d-9ad2-b12efb75e1aa\") " pod="openshift-marketplace/community-operators-4m4pm" Oct 01 09:40:33 crc kubenswrapper[4787]: I1001 09:40:33.056065 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e4be47d-87d5-469d-9ad2-b12efb75e1aa-utilities\") pod \"community-operators-4m4pm\" (UID: \"6e4be47d-87d5-469d-9ad2-b12efb75e1aa\") " pod="openshift-marketplace/community-operators-4m4pm" Oct 01 09:40:33 crc kubenswrapper[4787]: I1001 09:40:33.076558 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skt5t\" (UniqueName: \"kubernetes.io/projected/6e4be47d-87d5-469d-9ad2-b12efb75e1aa-kube-api-access-skt5t\") pod \"community-operators-4m4pm\" (UID: \"6e4be47d-87d5-469d-9ad2-b12efb75e1aa\") " pod="openshift-marketplace/community-operators-4m4pm" Oct 01 09:40:33 crc kubenswrapper[4787]: I1001 09:40:33.141663 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4m4pm" Oct 01 09:40:33 crc kubenswrapper[4787]: I1001 09:40:33.339726 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-swbhm" event={"ID":"593f4070-9419-4e2f-9da0-741e86877588","Type":"ContainerStarted","Data":"d59bc0d8f6fe55b1df7f1edb514125473af4a7c889df9e940861229eb1529ab1"} Oct 01 09:40:33 crc kubenswrapper[4787]: I1001 09:40:33.350454 4787 generic.go:334] "Generic (PLEG): container finished" podID="c3868e91-d922-4768-af6a-123795df5a28" containerID="8d0f3728cdd32b7b8f824f48ab5b5ce5f78722e04995e424875a96a8726f3e03" exitCode=0 Oct 01 09:40:33 crc kubenswrapper[4787]: I1001 09:40:33.350492 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pwj4z" event={"ID":"c3868e91-d922-4768-af6a-123795df5a28","Type":"ContainerDied","Data":"8d0f3728cdd32b7b8f824f48ab5b5ce5f78722e04995e424875a96a8726f3e03"} Oct 01 09:40:33 crc kubenswrapper[4787]: I1001 09:40:33.362988 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-swbhm" podStartSLOduration=1.937093535 podStartE2EDuration="3.362970516s" podCreationTimestamp="2025-10-01 09:40:30 +0000 UTC" firstStartedPulling="2025-10-01 09:40:31.315422235 +0000 UTC m=+263.430566392" lastFinishedPulling="2025-10-01 09:40:32.741299216 +0000 UTC m=+264.856443373" observedRunningTime="2025-10-01 09:40:33.359021883 +0000 UTC m=+265.474166050" watchObservedRunningTime="2025-10-01 09:40:33.362970516 +0000 UTC m=+265.478114673" Oct 01 09:40:33 crc kubenswrapper[4787]: I1001 09:40:33.421347 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8ksjp"] Oct 01 09:40:33 crc kubenswrapper[4787]: I1001 09:40:33.553066 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4m4pm"] Oct 01 09:40:33 crc kubenswrapper[4787]: W1001 09:40:33.560915 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e4be47d_87d5_469d_9ad2_b12efb75e1aa.slice/crio-c71285ccc8f58a6ed6564b38751f4240af31011cff161b08220e648ae3cca51f WatchSource:0}: Error finding container c71285ccc8f58a6ed6564b38751f4240af31011cff161b08220e648ae3cca51f: Status 404 returned error can't find the container with id c71285ccc8f58a6ed6564b38751f4240af31011cff161b08220e648ae3cca51f Oct 01 09:40:34 crc kubenswrapper[4787]: I1001 09:40:34.357179 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pwj4z" event={"ID":"c3868e91-d922-4768-af6a-123795df5a28","Type":"ContainerStarted","Data":"42002cb565d049a6cbeb24237d0296fe6d566af7ee22e0efce17d3d87b12d389"} Oct 01 09:40:34 crc kubenswrapper[4787]: I1001 09:40:34.360129 4787 generic.go:334] "Generic (PLEG): container finished" podID="6e4be47d-87d5-469d-9ad2-b12efb75e1aa" containerID="8c8899b4c39bd1341af6d06ce9ecdbc6973af2246fd41abebb5752ef8e37b030" exitCode=0 Oct 01 09:40:34 crc kubenswrapper[4787]: I1001 09:40:34.360189 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4m4pm" event={"ID":"6e4be47d-87d5-469d-9ad2-b12efb75e1aa","Type":"ContainerDied","Data":"8c8899b4c39bd1341af6d06ce9ecdbc6973af2246fd41abebb5752ef8e37b030"} Oct 01 09:40:34 crc kubenswrapper[4787]: I1001 09:40:34.360218 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4m4pm" event={"ID":"6e4be47d-87d5-469d-9ad2-b12efb75e1aa","Type":"ContainerStarted","Data":"c71285ccc8f58a6ed6564b38751f4240af31011cff161b08220e648ae3cca51f"} Oct 01 09:40:34 crc kubenswrapper[4787]: I1001 09:40:34.363807 4787 generic.go:334] "Generic (PLEG): container finished" podID="09d257f5-ca13-42c9-92f8-920f99ebadc1" containerID="696d1d63e0086b26f3d6433411513453fa4fd7158034fcd420970628ba766e07" exitCode=0 Oct 01 09:40:34 crc kubenswrapper[4787]: I1001 09:40:34.363882 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8ksjp" event={"ID":"09d257f5-ca13-42c9-92f8-920f99ebadc1","Type":"ContainerDied","Data":"696d1d63e0086b26f3d6433411513453fa4fd7158034fcd420970628ba766e07"} Oct 01 09:40:34 crc kubenswrapper[4787]: I1001 09:40:34.363913 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8ksjp" event={"ID":"09d257f5-ca13-42c9-92f8-920f99ebadc1","Type":"ContainerStarted","Data":"b0b8eb812cd0b288137883e9af32ae637526a7037b532c09f217c54545c076ae"} Oct 01 09:40:34 crc kubenswrapper[4787]: I1001 09:40:34.377882 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pwj4z" podStartSLOduration=1.845875807 podStartE2EDuration="4.377846848s" podCreationTimestamp="2025-10-01 09:40:30 +0000 UTC" firstStartedPulling="2025-10-01 09:40:31.312921924 +0000 UTC m=+263.428066091" lastFinishedPulling="2025-10-01 09:40:33.844892975 +0000 UTC m=+265.960037132" observedRunningTime="2025-10-01 09:40:34.37477947 +0000 UTC m=+266.489923617" watchObservedRunningTime="2025-10-01 09:40:34.377846848 +0000 UTC m=+266.492991005" Oct 01 09:40:37 crc kubenswrapper[4787]: I1001 09:40:37.381800 4787 generic.go:334] "Generic (PLEG): container finished" podID="6e4be47d-87d5-469d-9ad2-b12efb75e1aa" containerID="db94139e57e0fb1cf6e4852635ef0d2c0928baa882749eac5fab014069954247" exitCode=0 Oct 01 09:40:37 crc kubenswrapper[4787]: I1001 09:40:37.381879 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4m4pm" event={"ID":"6e4be47d-87d5-469d-9ad2-b12efb75e1aa","Type":"ContainerDied","Data":"db94139e57e0fb1cf6e4852635ef0d2c0928baa882749eac5fab014069954247"} Oct 01 09:40:37 crc kubenswrapper[4787]: I1001 09:40:37.388365 4787 generic.go:334] "Generic (PLEG): container finished" podID="09d257f5-ca13-42c9-92f8-920f99ebadc1" containerID="1916d6081f5aefd775304ee03aa87e0487972d405aa4766f75f20bf9f8b783de" exitCode=0 Oct 01 09:40:37 crc kubenswrapper[4787]: I1001 09:40:37.388400 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8ksjp" event={"ID":"09d257f5-ca13-42c9-92f8-920f99ebadc1","Type":"ContainerDied","Data":"1916d6081f5aefd775304ee03aa87e0487972d405aa4766f75f20bf9f8b783de"} Oct 01 09:40:38 crc kubenswrapper[4787]: I1001 09:40:38.398392 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4m4pm" event={"ID":"6e4be47d-87d5-469d-9ad2-b12efb75e1aa","Type":"ContainerStarted","Data":"b52d86a930ccb6b5bc0668716b9eda3d4180a1776dbd6aa936cc313f770b8964"} Oct 01 09:40:38 crc kubenswrapper[4787]: I1001 09:40:38.400864 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8ksjp" event={"ID":"09d257f5-ca13-42c9-92f8-920f99ebadc1","Type":"ContainerStarted","Data":"a5cd7e28c8d213d0fc0f089e1f7d5246004a434a8f33dd5e57e97ce69168af20"} Oct 01 09:40:38 crc kubenswrapper[4787]: I1001 09:40:38.416765 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4m4pm" podStartSLOduration=2.732659726 podStartE2EDuration="6.416742617s" podCreationTimestamp="2025-10-01 09:40:32 +0000 UTC" firstStartedPulling="2025-10-01 09:40:34.361529783 +0000 UTC m=+266.476673940" lastFinishedPulling="2025-10-01 09:40:38.045612674 +0000 UTC m=+270.160756831" observedRunningTime="2025-10-01 09:40:38.414961487 +0000 UTC m=+270.530105684" watchObservedRunningTime="2025-10-01 09:40:38.416742617 +0000 UTC m=+270.531886774" Oct 01 09:40:38 crc kubenswrapper[4787]: I1001 09:40:38.438420 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8ksjp" podStartSLOduration=2.654201992 podStartE2EDuration="6.438396835s" podCreationTimestamp="2025-10-01 09:40:32 +0000 UTC" firstStartedPulling="2025-10-01 09:40:34.366930997 +0000 UTC m=+266.482075154" lastFinishedPulling="2025-10-01 09:40:38.15112584 +0000 UTC m=+270.266269997" observedRunningTime="2025-10-01 09:40:38.43577961 +0000 UTC m=+270.550923767" watchObservedRunningTime="2025-10-01 09:40:38.438396835 +0000 UTC m=+270.553540992" Oct 01 09:40:40 crc kubenswrapper[4787]: I1001 09:40:40.545357 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-swbhm" Oct 01 09:40:40 crc kubenswrapper[4787]: I1001 09:40:40.545914 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-swbhm" Oct 01 09:40:40 crc kubenswrapper[4787]: I1001 09:40:40.589979 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-swbhm" Oct 01 09:40:40 crc kubenswrapper[4787]: I1001 09:40:40.761907 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pwj4z" Oct 01 09:40:40 crc kubenswrapper[4787]: I1001 09:40:40.761948 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pwj4z" Oct 01 09:40:40 crc kubenswrapper[4787]: I1001 09:40:40.797332 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pwj4z" Oct 01 09:40:41 crc kubenswrapper[4787]: I1001 09:40:41.453773 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-swbhm" Oct 01 09:40:41 crc kubenswrapper[4787]: I1001 09:40:41.476285 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pwj4z" Oct 01 09:40:42 crc kubenswrapper[4787]: I1001 09:40:42.995931 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8ksjp" Oct 01 09:40:42 crc kubenswrapper[4787]: I1001 09:40:42.995985 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8ksjp" Oct 01 09:40:43 crc kubenswrapper[4787]: I1001 09:40:43.049254 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8ksjp" Oct 01 09:40:43 crc kubenswrapper[4787]: I1001 09:40:43.142734 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4m4pm" Oct 01 09:40:43 crc kubenswrapper[4787]: I1001 09:40:43.142784 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4m4pm" Oct 01 09:40:43 crc kubenswrapper[4787]: I1001 09:40:43.176319 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4m4pm" Oct 01 09:40:43 crc kubenswrapper[4787]: I1001 09:40:43.467396 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8ksjp" Oct 01 09:40:43 crc kubenswrapper[4787]: I1001 09:40:43.480693 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4m4pm" Oct 01 09:41:41 crc kubenswrapper[4787]: I1001 09:41:41.250786 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:41:41 crc kubenswrapper[4787]: I1001 09:41:41.251478 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:42:11 crc kubenswrapper[4787]: I1001 09:42:11.250328 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:42:11 crc kubenswrapper[4787]: I1001 09:42:11.250912 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:42:41 crc kubenswrapper[4787]: I1001 09:42:41.251286 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:42:41 crc kubenswrapper[4787]: I1001 09:42:41.251878 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:42:41 crc kubenswrapper[4787]: I1001 09:42:41.251932 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" Oct 01 09:42:41 crc kubenswrapper[4787]: I1001 09:42:41.252602 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bc6fdce923b6e1a3fe85923c737b807469b5f84a3dceb7d7e024941e22a498f2"} pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 09:42:41 crc kubenswrapper[4787]: I1001 09:42:41.252675 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" containerID="cri-o://bc6fdce923b6e1a3fe85923c737b807469b5f84a3dceb7d7e024941e22a498f2" gracePeriod=600 Oct 01 09:42:42 crc kubenswrapper[4787]: I1001 09:42:42.101591 4787 generic.go:334] "Generic (PLEG): container finished" podID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerID="bc6fdce923b6e1a3fe85923c737b807469b5f84a3dceb7d7e024941e22a498f2" exitCode=0 Oct 01 09:42:42 crc kubenswrapper[4787]: I1001 09:42:42.101724 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" event={"ID":"fa51b883-5f3e-4141-9d4c-37704aac7718","Type":"ContainerDied","Data":"bc6fdce923b6e1a3fe85923c737b807469b5f84a3dceb7d7e024941e22a498f2"} Oct 01 09:42:42 crc kubenswrapper[4787]: I1001 09:42:42.101978 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" event={"ID":"fa51b883-5f3e-4141-9d4c-37704aac7718","Type":"ContainerStarted","Data":"e4d458228b35af45ba3f97cae60443c240914647056563a211cb5a9e256ce4f6"} Oct 01 09:42:42 crc kubenswrapper[4787]: I1001 09:42:42.102012 4787 scope.go:117] "RemoveContainer" containerID="50f44e85c30d1925922640deb44aa669132a1e7d41e948d593cb863bb0dd6188" Oct 01 09:43:47 crc kubenswrapper[4787]: I1001 09:43:47.383998 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-fh57g"] Oct 01 09:43:47 crc kubenswrapper[4787]: I1001 09:43:47.385597 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-fh57g" Oct 01 09:43:47 crc kubenswrapper[4787]: I1001 09:43:47.411909 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-fh57g"] Oct 01 09:43:47 crc kubenswrapper[4787]: I1001 09:43:47.534442 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/899c737a-3bed-48fd-ab97-f6445d0e18dc-trusted-ca\") pod \"image-registry-66df7c8f76-fh57g\" (UID: \"899c737a-3bed-48fd-ab97-f6445d0e18dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-fh57g" Oct 01 09:43:47 crc kubenswrapper[4787]: I1001 09:43:47.534515 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/899c737a-3bed-48fd-ab97-f6445d0e18dc-registry-tls\") pod \"image-registry-66df7c8f76-fh57g\" (UID: \"899c737a-3bed-48fd-ab97-f6445d0e18dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-fh57g" Oct 01 09:43:47 crc kubenswrapper[4787]: I1001 09:43:47.534562 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tq2d8\" (UniqueName: \"kubernetes.io/projected/899c737a-3bed-48fd-ab97-f6445d0e18dc-kube-api-access-tq2d8\") pod \"image-registry-66df7c8f76-fh57g\" (UID: \"899c737a-3bed-48fd-ab97-f6445d0e18dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-fh57g" Oct 01 09:43:47 crc kubenswrapper[4787]: I1001 09:43:47.534592 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/899c737a-3bed-48fd-ab97-f6445d0e18dc-registry-certificates\") pod \"image-registry-66df7c8f76-fh57g\" (UID: \"899c737a-3bed-48fd-ab97-f6445d0e18dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-fh57g" Oct 01 09:43:47 crc kubenswrapper[4787]: I1001 09:43:47.534622 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/899c737a-3bed-48fd-ab97-f6445d0e18dc-installation-pull-secrets\") pod \"image-registry-66df7c8f76-fh57g\" (UID: \"899c737a-3bed-48fd-ab97-f6445d0e18dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-fh57g" Oct 01 09:43:47 crc kubenswrapper[4787]: I1001 09:43:47.534649 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/899c737a-3bed-48fd-ab97-f6445d0e18dc-ca-trust-extracted\") pod \"image-registry-66df7c8f76-fh57g\" (UID: \"899c737a-3bed-48fd-ab97-f6445d0e18dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-fh57g" Oct 01 09:43:47 crc kubenswrapper[4787]: I1001 09:43:47.534676 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/899c737a-3bed-48fd-ab97-f6445d0e18dc-bound-sa-token\") pod \"image-registry-66df7c8f76-fh57g\" (UID: \"899c737a-3bed-48fd-ab97-f6445d0e18dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-fh57g" Oct 01 09:43:47 crc kubenswrapper[4787]: I1001 09:43:47.534775 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-fh57g\" (UID: \"899c737a-3bed-48fd-ab97-f6445d0e18dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-fh57g" Oct 01 09:43:47 crc kubenswrapper[4787]: I1001 09:43:47.563991 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-fh57g\" (UID: \"899c737a-3bed-48fd-ab97-f6445d0e18dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-fh57g" Oct 01 09:43:47 crc kubenswrapper[4787]: I1001 09:43:47.636426 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tq2d8\" (UniqueName: \"kubernetes.io/projected/899c737a-3bed-48fd-ab97-f6445d0e18dc-kube-api-access-tq2d8\") pod \"image-registry-66df7c8f76-fh57g\" (UID: \"899c737a-3bed-48fd-ab97-f6445d0e18dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-fh57g" Oct 01 09:43:47 crc kubenswrapper[4787]: I1001 09:43:47.636773 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/899c737a-3bed-48fd-ab97-f6445d0e18dc-registry-certificates\") pod \"image-registry-66df7c8f76-fh57g\" (UID: \"899c737a-3bed-48fd-ab97-f6445d0e18dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-fh57g" Oct 01 09:43:47 crc kubenswrapper[4787]: I1001 09:43:47.636899 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/899c737a-3bed-48fd-ab97-f6445d0e18dc-installation-pull-secrets\") pod \"image-registry-66df7c8f76-fh57g\" (UID: \"899c737a-3bed-48fd-ab97-f6445d0e18dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-fh57g" Oct 01 09:43:47 crc kubenswrapper[4787]: I1001 09:43:47.637031 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/899c737a-3bed-48fd-ab97-f6445d0e18dc-ca-trust-extracted\") pod \"image-registry-66df7c8f76-fh57g\" (UID: \"899c737a-3bed-48fd-ab97-f6445d0e18dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-fh57g" Oct 01 09:43:47 crc kubenswrapper[4787]: I1001 09:43:47.637168 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/899c737a-3bed-48fd-ab97-f6445d0e18dc-bound-sa-token\") pod \"image-registry-66df7c8f76-fh57g\" (UID: \"899c737a-3bed-48fd-ab97-f6445d0e18dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-fh57g" Oct 01 09:43:47 crc kubenswrapper[4787]: I1001 09:43:47.637316 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/899c737a-3bed-48fd-ab97-f6445d0e18dc-trusted-ca\") pod \"image-registry-66df7c8f76-fh57g\" (UID: \"899c737a-3bed-48fd-ab97-f6445d0e18dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-fh57g" Oct 01 09:43:47 crc kubenswrapper[4787]: I1001 09:43:47.637444 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/899c737a-3bed-48fd-ab97-f6445d0e18dc-registry-tls\") pod \"image-registry-66df7c8f76-fh57g\" (UID: \"899c737a-3bed-48fd-ab97-f6445d0e18dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-fh57g" Oct 01 09:43:47 crc kubenswrapper[4787]: I1001 09:43:47.637560 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/899c737a-3bed-48fd-ab97-f6445d0e18dc-ca-trust-extracted\") pod \"image-registry-66df7c8f76-fh57g\" (UID: \"899c737a-3bed-48fd-ab97-f6445d0e18dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-fh57g" Oct 01 09:43:47 crc kubenswrapper[4787]: I1001 09:43:47.639913 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/899c737a-3bed-48fd-ab97-f6445d0e18dc-trusted-ca\") pod \"image-registry-66df7c8f76-fh57g\" (UID: \"899c737a-3bed-48fd-ab97-f6445d0e18dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-fh57g" Oct 01 09:43:47 crc kubenswrapper[4787]: I1001 09:43:47.640862 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/899c737a-3bed-48fd-ab97-f6445d0e18dc-registry-certificates\") pod \"image-registry-66df7c8f76-fh57g\" (UID: \"899c737a-3bed-48fd-ab97-f6445d0e18dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-fh57g" Oct 01 09:43:47 crc kubenswrapper[4787]: I1001 09:43:47.644419 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/899c737a-3bed-48fd-ab97-f6445d0e18dc-installation-pull-secrets\") pod \"image-registry-66df7c8f76-fh57g\" (UID: \"899c737a-3bed-48fd-ab97-f6445d0e18dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-fh57g" Oct 01 09:43:47 crc kubenswrapper[4787]: I1001 09:43:47.644575 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/899c737a-3bed-48fd-ab97-f6445d0e18dc-registry-tls\") pod \"image-registry-66df7c8f76-fh57g\" (UID: \"899c737a-3bed-48fd-ab97-f6445d0e18dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-fh57g" Oct 01 09:43:47 crc kubenswrapper[4787]: I1001 09:43:47.651977 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tq2d8\" (UniqueName: \"kubernetes.io/projected/899c737a-3bed-48fd-ab97-f6445d0e18dc-kube-api-access-tq2d8\") pod \"image-registry-66df7c8f76-fh57g\" (UID: \"899c737a-3bed-48fd-ab97-f6445d0e18dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-fh57g" Oct 01 09:43:47 crc kubenswrapper[4787]: I1001 09:43:47.652656 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/899c737a-3bed-48fd-ab97-f6445d0e18dc-bound-sa-token\") pod \"image-registry-66df7c8f76-fh57g\" (UID: \"899c737a-3bed-48fd-ab97-f6445d0e18dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-fh57g" Oct 01 09:43:47 crc kubenswrapper[4787]: I1001 09:43:47.707261 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-fh57g" Oct 01 09:43:48 crc kubenswrapper[4787]: I1001 09:43:48.117112 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-fh57g"] Oct 01 09:43:48 crc kubenswrapper[4787]: I1001 09:43:48.531334 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-fh57g" event={"ID":"899c737a-3bed-48fd-ab97-f6445d0e18dc","Type":"ContainerStarted","Data":"6ba1847da1e07ed6454805b8f3223dea2cda703de602b1e7eaf769f8fa447939"} Oct 01 09:43:48 crc kubenswrapper[4787]: I1001 09:43:48.533111 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-fh57g" event={"ID":"899c737a-3bed-48fd-ab97-f6445d0e18dc","Type":"ContainerStarted","Data":"9765ecdabc329abc08b7b0d2b34071d163a586ed96cd3ac68511711d13228d78"} Oct 01 09:43:48 crc kubenswrapper[4787]: I1001 09:43:48.533254 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-fh57g" Oct 01 09:43:48 crc kubenswrapper[4787]: I1001 09:43:48.564184 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-fh57g" podStartSLOduration=1.564161258 podStartE2EDuration="1.564161258s" podCreationTimestamp="2025-10-01 09:43:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:43:48.559544936 +0000 UTC m=+460.674689113" watchObservedRunningTime="2025-10-01 09:43:48.564161258 +0000 UTC m=+460.679305435" Oct 01 09:44:07 crc kubenswrapper[4787]: I1001 09:44:07.712527 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-fh57g" Oct 01 09:44:07 crc kubenswrapper[4787]: I1001 09:44:07.778668 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cpqnn"] Oct 01 09:44:32 crc kubenswrapper[4787]: I1001 09:44:32.831402 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" podUID="8ffe49e7-9382-4711-985c-8d6f76c42abc" containerName="registry" containerID="cri-o://24e6211e9d0aba2dcda0020a7110833544f47433f0c35f6d9ce9aad377589a16" gracePeriod=30 Oct 01 09:44:33 crc kubenswrapper[4787]: I1001 09:44:33.184153 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:44:33 crc kubenswrapper[4787]: I1001 09:44:33.368393 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8ffe49e7-9382-4711-985c-8d6f76c42abc\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " Oct 01 09:44:33 crc kubenswrapper[4787]: I1001 09:44:33.368533 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgmrq\" (UniqueName: \"kubernetes.io/projected/8ffe49e7-9382-4711-985c-8d6f76c42abc-kube-api-access-dgmrq\") pod \"8ffe49e7-9382-4711-985c-8d6f76c42abc\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " Oct 01 09:44:33 crc kubenswrapper[4787]: I1001 09:44:33.368566 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8ffe49e7-9382-4711-985c-8d6f76c42abc-trusted-ca\") pod \"8ffe49e7-9382-4711-985c-8d6f76c42abc\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " Oct 01 09:44:33 crc kubenswrapper[4787]: I1001 09:44:33.368618 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8ffe49e7-9382-4711-985c-8d6f76c42abc-registry-certificates\") pod \"8ffe49e7-9382-4711-985c-8d6f76c42abc\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " Oct 01 09:44:33 crc kubenswrapper[4787]: I1001 09:44:33.368654 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8ffe49e7-9382-4711-985c-8d6f76c42abc-registry-tls\") pod \"8ffe49e7-9382-4711-985c-8d6f76c42abc\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " Oct 01 09:44:33 crc kubenswrapper[4787]: I1001 09:44:33.368704 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8ffe49e7-9382-4711-985c-8d6f76c42abc-installation-pull-secrets\") pod \"8ffe49e7-9382-4711-985c-8d6f76c42abc\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " Oct 01 09:44:33 crc kubenswrapper[4787]: I1001 09:44:33.368748 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8ffe49e7-9382-4711-985c-8d6f76c42abc-ca-trust-extracted\") pod \"8ffe49e7-9382-4711-985c-8d6f76c42abc\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " Oct 01 09:44:33 crc kubenswrapper[4787]: I1001 09:44:33.368790 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8ffe49e7-9382-4711-985c-8d6f76c42abc-bound-sa-token\") pod \"8ffe49e7-9382-4711-985c-8d6f76c42abc\" (UID: \"8ffe49e7-9382-4711-985c-8d6f76c42abc\") " Oct 01 09:44:33 crc kubenswrapper[4787]: I1001 09:44:33.369805 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ffe49e7-9382-4711-985c-8d6f76c42abc-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8ffe49e7-9382-4711-985c-8d6f76c42abc" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:44:33 crc kubenswrapper[4787]: I1001 09:44:33.369932 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ffe49e7-9382-4711-985c-8d6f76c42abc-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8ffe49e7-9382-4711-985c-8d6f76c42abc" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:44:33 crc kubenswrapper[4787]: I1001 09:44:33.374618 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ffe49e7-9382-4711-985c-8d6f76c42abc-kube-api-access-dgmrq" (OuterVolumeSpecName: "kube-api-access-dgmrq") pod "8ffe49e7-9382-4711-985c-8d6f76c42abc" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc"). InnerVolumeSpecName "kube-api-access-dgmrq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:44:33 crc kubenswrapper[4787]: I1001 09:44:33.376341 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ffe49e7-9382-4711-985c-8d6f76c42abc-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8ffe49e7-9382-4711-985c-8d6f76c42abc" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:44:33 crc kubenswrapper[4787]: I1001 09:44:33.376686 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ffe49e7-9382-4711-985c-8d6f76c42abc-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8ffe49e7-9382-4711-985c-8d6f76c42abc" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:44:33 crc kubenswrapper[4787]: I1001 09:44:33.377193 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ffe49e7-9382-4711-985c-8d6f76c42abc-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8ffe49e7-9382-4711-985c-8d6f76c42abc" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:44:33 crc kubenswrapper[4787]: I1001 09:44:33.379480 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "8ffe49e7-9382-4711-985c-8d6f76c42abc" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 01 09:44:33 crc kubenswrapper[4787]: I1001 09:44:33.386174 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ffe49e7-9382-4711-985c-8d6f76c42abc-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8ffe49e7-9382-4711-985c-8d6f76c42abc" (UID: "8ffe49e7-9382-4711-985c-8d6f76c42abc"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:44:33 crc kubenswrapper[4787]: I1001 09:44:33.469984 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgmrq\" (UniqueName: \"kubernetes.io/projected/8ffe49e7-9382-4711-985c-8d6f76c42abc-kube-api-access-dgmrq\") on node \"crc\" DevicePath \"\"" Oct 01 09:44:33 crc kubenswrapper[4787]: I1001 09:44:33.470064 4787 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8ffe49e7-9382-4711-985c-8d6f76c42abc-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:44:33 crc kubenswrapper[4787]: I1001 09:44:33.470111 4787 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8ffe49e7-9382-4711-985c-8d6f76c42abc-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 01 09:44:33 crc kubenswrapper[4787]: I1001 09:44:33.470126 4787 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8ffe49e7-9382-4711-985c-8d6f76c42abc-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 01 09:44:33 crc kubenswrapper[4787]: I1001 09:44:33.470138 4787 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8ffe49e7-9382-4711-985c-8d6f76c42abc-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 01 09:44:33 crc kubenswrapper[4787]: I1001 09:44:33.470149 4787 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8ffe49e7-9382-4711-985c-8d6f76c42abc-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 01 09:44:33 crc kubenswrapper[4787]: I1001 09:44:33.470165 4787 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8ffe49e7-9382-4711-985c-8d6f76c42abc-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 01 09:44:33 crc kubenswrapper[4787]: I1001 09:44:33.820463 4787 generic.go:334] "Generic (PLEG): container finished" podID="8ffe49e7-9382-4711-985c-8d6f76c42abc" containerID="24e6211e9d0aba2dcda0020a7110833544f47433f0c35f6d9ce9aad377589a16" exitCode=0 Oct 01 09:44:33 crc kubenswrapper[4787]: I1001 09:44:33.820511 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" event={"ID":"8ffe49e7-9382-4711-985c-8d6f76c42abc","Type":"ContainerDied","Data":"24e6211e9d0aba2dcda0020a7110833544f47433f0c35f6d9ce9aad377589a16"} Oct 01 09:44:33 crc kubenswrapper[4787]: I1001 09:44:33.820543 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" event={"ID":"8ffe49e7-9382-4711-985c-8d6f76c42abc","Type":"ContainerDied","Data":"924be146bfef052c7afcc388ecd46b6c52f34b9ef3c62a14b479dc715baca166"} Oct 01 09:44:33 crc kubenswrapper[4787]: I1001 09:44:33.820549 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-cpqnn" Oct 01 09:44:33 crc kubenswrapper[4787]: I1001 09:44:33.820559 4787 scope.go:117] "RemoveContainer" containerID="24e6211e9d0aba2dcda0020a7110833544f47433f0c35f6d9ce9aad377589a16" Oct 01 09:44:33 crc kubenswrapper[4787]: I1001 09:44:33.841117 4787 scope.go:117] "RemoveContainer" containerID="24e6211e9d0aba2dcda0020a7110833544f47433f0c35f6d9ce9aad377589a16" Oct 01 09:44:33 crc kubenswrapper[4787]: E1001 09:44:33.842032 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24e6211e9d0aba2dcda0020a7110833544f47433f0c35f6d9ce9aad377589a16\": container with ID starting with 24e6211e9d0aba2dcda0020a7110833544f47433f0c35f6d9ce9aad377589a16 not found: ID does not exist" containerID="24e6211e9d0aba2dcda0020a7110833544f47433f0c35f6d9ce9aad377589a16" Oct 01 09:44:33 crc kubenswrapper[4787]: I1001 09:44:33.842146 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24e6211e9d0aba2dcda0020a7110833544f47433f0c35f6d9ce9aad377589a16"} err="failed to get container status \"24e6211e9d0aba2dcda0020a7110833544f47433f0c35f6d9ce9aad377589a16\": rpc error: code = NotFound desc = could not find container \"24e6211e9d0aba2dcda0020a7110833544f47433f0c35f6d9ce9aad377589a16\": container with ID starting with 24e6211e9d0aba2dcda0020a7110833544f47433f0c35f6d9ce9aad377589a16 not found: ID does not exist" Oct 01 09:44:33 crc kubenswrapper[4787]: I1001 09:44:33.852714 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cpqnn"] Oct 01 09:44:33 crc kubenswrapper[4787]: I1001 09:44:33.858515 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cpqnn"] Oct 01 09:44:34 crc kubenswrapper[4787]: I1001 09:44:34.531896 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ffe49e7-9382-4711-985c-8d6f76c42abc" path="/var/lib/kubelet/pods/8ffe49e7-9382-4711-985c-8d6f76c42abc/volumes" Oct 01 09:44:41 crc kubenswrapper[4787]: I1001 09:44:41.250805 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:44:41 crc kubenswrapper[4787]: I1001 09:44:41.251260 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:45:00 crc kubenswrapper[4787]: I1001 09:45:00.145665 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321865-9nl5h"] Oct 01 09:45:00 crc kubenswrapper[4787]: E1001 09:45:00.150050 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ffe49e7-9382-4711-985c-8d6f76c42abc" containerName="registry" Oct 01 09:45:00 crc kubenswrapper[4787]: I1001 09:45:00.150102 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ffe49e7-9382-4711-985c-8d6f76c42abc" containerName="registry" Oct 01 09:45:00 crc kubenswrapper[4787]: I1001 09:45:00.150263 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ffe49e7-9382-4711-985c-8d6f76c42abc" containerName="registry" Oct 01 09:45:00 crc kubenswrapper[4787]: I1001 09:45:00.150772 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321865-9nl5h" Oct 01 09:45:00 crc kubenswrapper[4787]: I1001 09:45:00.152872 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 01 09:45:00 crc kubenswrapper[4787]: I1001 09:45:00.153419 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 01 09:45:00 crc kubenswrapper[4787]: I1001 09:45:00.154599 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321865-9nl5h"] Oct 01 09:45:00 crc kubenswrapper[4787]: I1001 09:45:00.234371 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c159d29b-be19-4d0f-980e-a12c2ed641f2-config-volume\") pod \"collect-profiles-29321865-9nl5h\" (UID: \"c159d29b-be19-4d0f-980e-a12c2ed641f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321865-9nl5h" Oct 01 09:45:00 crc kubenswrapper[4787]: I1001 09:45:00.234836 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c159d29b-be19-4d0f-980e-a12c2ed641f2-secret-volume\") pod \"collect-profiles-29321865-9nl5h\" (UID: \"c159d29b-be19-4d0f-980e-a12c2ed641f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321865-9nl5h" Oct 01 09:45:00 crc kubenswrapper[4787]: I1001 09:45:00.234966 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsf4q\" (UniqueName: \"kubernetes.io/projected/c159d29b-be19-4d0f-980e-a12c2ed641f2-kube-api-access-jsf4q\") pod \"collect-profiles-29321865-9nl5h\" (UID: \"c159d29b-be19-4d0f-980e-a12c2ed641f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321865-9nl5h" Oct 01 09:45:00 crc kubenswrapper[4787]: I1001 09:45:00.336439 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c159d29b-be19-4d0f-980e-a12c2ed641f2-config-volume\") pod \"collect-profiles-29321865-9nl5h\" (UID: \"c159d29b-be19-4d0f-980e-a12c2ed641f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321865-9nl5h" Oct 01 09:45:00 crc kubenswrapper[4787]: I1001 09:45:00.336785 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c159d29b-be19-4d0f-980e-a12c2ed641f2-secret-volume\") pod \"collect-profiles-29321865-9nl5h\" (UID: \"c159d29b-be19-4d0f-980e-a12c2ed641f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321865-9nl5h" Oct 01 09:45:00 crc kubenswrapper[4787]: I1001 09:45:00.336981 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsf4q\" (UniqueName: \"kubernetes.io/projected/c159d29b-be19-4d0f-980e-a12c2ed641f2-kube-api-access-jsf4q\") pod \"collect-profiles-29321865-9nl5h\" (UID: \"c159d29b-be19-4d0f-980e-a12c2ed641f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321865-9nl5h" Oct 01 09:45:00 crc kubenswrapper[4787]: I1001 09:45:00.337424 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c159d29b-be19-4d0f-980e-a12c2ed641f2-config-volume\") pod \"collect-profiles-29321865-9nl5h\" (UID: \"c159d29b-be19-4d0f-980e-a12c2ed641f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321865-9nl5h" Oct 01 09:45:00 crc kubenswrapper[4787]: I1001 09:45:00.342147 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c159d29b-be19-4d0f-980e-a12c2ed641f2-secret-volume\") pod \"collect-profiles-29321865-9nl5h\" (UID: \"c159d29b-be19-4d0f-980e-a12c2ed641f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321865-9nl5h" Oct 01 09:45:00 crc kubenswrapper[4787]: I1001 09:45:00.357007 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsf4q\" (UniqueName: \"kubernetes.io/projected/c159d29b-be19-4d0f-980e-a12c2ed641f2-kube-api-access-jsf4q\") pod \"collect-profiles-29321865-9nl5h\" (UID: \"c159d29b-be19-4d0f-980e-a12c2ed641f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321865-9nl5h" Oct 01 09:45:00 crc kubenswrapper[4787]: I1001 09:45:00.478913 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321865-9nl5h" Oct 01 09:45:00 crc kubenswrapper[4787]: I1001 09:45:00.678846 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321865-9nl5h"] Oct 01 09:45:00 crc kubenswrapper[4787]: W1001 09:45:00.686979 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc159d29b_be19_4d0f_980e_a12c2ed641f2.slice/crio-04816764c8cc050999e983d29d9032437488ae571b188874da4cb72143bd6ce7 WatchSource:0}: Error finding container 04816764c8cc050999e983d29d9032437488ae571b188874da4cb72143bd6ce7: Status 404 returned error can't find the container with id 04816764c8cc050999e983d29d9032437488ae571b188874da4cb72143bd6ce7 Oct 01 09:45:00 crc kubenswrapper[4787]: I1001 09:45:00.980997 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321865-9nl5h" event={"ID":"c159d29b-be19-4d0f-980e-a12c2ed641f2","Type":"ContainerStarted","Data":"863b0a8fe30dcfd7f3f0f2313758f642a66355859da1809cf9417915270239f0"} Oct 01 09:45:00 crc kubenswrapper[4787]: I1001 09:45:00.981057 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321865-9nl5h" event={"ID":"c159d29b-be19-4d0f-980e-a12c2ed641f2","Type":"ContainerStarted","Data":"04816764c8cc050999e983d29d9032437488ae571b188874da4cb72143bd6ce7"} Oct 01 09:45:00 crc kubenswrapper[4787]: I1001 09:45:00.994835 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29321865-9nl5h" podStartSLOduration=0.994804768 podStartE2EDuration="994.804768ms" podCreationTimestamp="2025-10-01 09:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:45:00.99452637 +0000 UTC m=+533.109670537" watchObservedRunningTime="2025-10-01 09:45:00.994804768 +0000 UTC m=+533.109948965" Oct 01 09:45:01 crc kubenswrapper[4787]: I1001 09:45:01.989869 4787 generic.go:334] "Generic (PLEG): container finished" podID="c159d29b-be19-4d0f-980e-a12c2ed641f2" containerID="863b0a8fe30dcfd7f3f0f2313758f642a66355859da1809cf9417915270239f0" exitCode=0 Oct 01 09:45:01 crc kubenswrapper[4787]: I1001 09:45:01.989976 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321865-9nl5h" event={"ID":"c159d29b-be19-4d0f-980e-a12c2ed641f2","Type":"ContainerDied","Data":"863b0a8fe30dcfd7f3f0f2313758f642a66355859da1809cf9417915270239f0"} Oct 01 09:45:03 crc kubenswrapper[4787]: I1001 09:45:03.214731 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321865-9nl5h" Oct 01 09:45:03 crc kubenswrapper[4787]: I1001 09:45:03.377596 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jsf4q\" (UniqueName: \"kubernetes.io/projected/c159d29b-be19-4d0f-980e-a12c2ed641f2-kube-api-access-jsf4q\") pod \"c159d29b-be19-4d0f-980e-a12c2ed641f2\" (UID: \"c159d29b-be19-4d0f-980e-a12c2ed641f2\") " Oct 01 09:45:03 crc kubenswrapper[4787]: I1001 09:45:03.377727 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c159d29b-be19-4d0f-980e-a12c2ed641f2-secret-volume\") pod \"c159d29b-be19-4d0f-980e-a12c2ed641f2\" (UID: \"c159d29b-be19-4d0f-980e-a12c2ed641f2\") " Oct 01 09:45:03 crc kubenswrapper[4787]: I1001 09:45:03.377811 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c159d29b-be19-4d0f-980e-a12c2ed641f2-config-volume\") pod \"c159d29b-be19-4d0f-980e-a12c2ed641f2\" (UID: \"c159d29b-be19-4d0f-980e-a12c2ed641f2\") " Oct 01 09:45:03 crc kubenswrapper[4787]: I1001 09:45:03.378691 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c159d29b-be19-4d0f-980e-a12c2ed641f2-config-volume" (OuterVolumeSpecName: "config-volume") pod "c159d29b-be19-4d0f-980e-a12c2ed641f2" (UID: "c159d29b-be19-4d0f-980e-a12c2ed641f2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:45:03 crc kubenswrapper[4787]: I1001 09:45:03.383755 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c159d29b-be19-4d0f-980e-a12c2ed641f2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c159d29b-be19-4d0f-980e-a12c2ed641f2" (UID: "c159d29b-be19-4d0f-980e-a12c2ed641f2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:45:03 crc kubenswrapper[4787]: I1001 09:45:03.383821 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c159d29b-be19-4d0f-980e-a12c2ed641f2-kube-api-access-jsf4q" (OuterVolumeSpecName: "kube-api-access-jsf4q") pod "c159d29b-be19-4d0f-980e-a12c2ed641f2" (UID: "c159d29b-be19-4d0f-980e-a12c2ed641f2"). InnerVolumeSpecName "kube-api-access-jsf4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:45:03 crc kubenswrapper[4787]: I1001 09:45:03.478888 4787 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c159d29b-be19-4d0f-980e-a12c2ed641f2-config-volume\") on node \"crc\" DevicePath \"\"" Oct 01 09:45:03 crc kubenswrapper[4787]: I1001 09:45:03.478924 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jsf4q\" (UniqueName: \"kubernetes.io/projected/c159d29b-be19-4d0f-980e-a12c2ed641f2-kube-api-access-jsf4q\") on node \"crc\" DevicePath \"\"" Oct 01 09:45:03 crc kubenswrapper[4787]: I1001 09:45:03.478936 4787 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c159d29b-be19-4d0f-980e-a12c2ed641f2-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 01 09:45:04 crc kubenswrapper[4787]: I1001 09:45:04.001245 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321865-9nl5h" event={"ID":"c159d29b-be19-4d0f-980e-a12c2ed641f2","Type":"ContainerDied","Data":"04816764c8cc050999e983d29d9032437488ae571b188874da4cb72143bd6ce7"} Oct 01 09:45:04 crc kubenswrapper[4787]: I1001 09:45:04.001306 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04816764c8cc050999e983d29d9032437488ae571b188874da4cb72143bd6ce7" Oct 01 09:45:04 crc kubenswrapper[4787]: I1001 09:45:04.001357 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321865-9nl5h" Oct 01 09:45:11 crc kubenswrapper[4787]: I1001 09:45:11.250859 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:45:11 crc kubenswrapper[4787]: I1001 09:45:11.252297 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:45:41 crc kubenswrapper[4787]: I1001 09:45:41.250495 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:45:41 crc kubenswrapper[4787]: I1001 09:45:41.251212 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:45:41 crc kubenswrapper[4787]: I1001 09:45:41.251271 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" Oct 01 09:45:41 crc kubenswrapper[4787]: I1001 09:45:41.252105 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e4d458228b35af45ba3f97cae60443c240914647056563a211cb5a9e256ce4f6"} pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 09:45:41 crc kubenswrapper[4787]: I1001 09:45:41.252216 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" containerID="cri-o://e4d458228b35af45ba3f97cae60443c240914647056563a211cb5a9e256ce4f6" gracePeriod=600 Oct 01 09:45:42 crc kubenswrapper[4787]: I1001 09:45:42.197200 4787 generic.go:334] "Generic (PLEG): container finished" podID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerID="e4d458228b35af45ba3f97cae60443c240914647056563a211cb5a9e256ce4f6" exitCode=0 Oct 01 09:45:42 crc kubenswrapper[4787]: I1001 09:45:42.197296 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" event={"ID":"fa51b883-5f3e-4141-9d4c-37704aac7718","Type":"ContainerDied","Data":"e4d458228b35af45ba3f97cae60443c240914647056563a211cb5a9e256ce4f6"} Oct 01 09:45:42 crc kubenswrapper[4787]: I1001 09:45:42.197777 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" event={"ID":"fa51b883-5f3e-4141-9d4c-37704aac7718","Type":"ContainerStarted","Data":"c5956c6b3dfb9dae0b884bab2812d62e85ed0b6d8154a894d187e5889824b51f"} Oct 01 09:45:42 crc kubenswrapper[4787]: I1001 09:45:42.197818 4787 scope.go:117] "RemoveContainer" containerID="bc6fdce923b6e1a3fe85923c737b807469b5f84a3dceb7d7e024941e22a498f2" Oct 01 09:45:42 crc kubenswrapper[4787]: I1001 09:45:42.568139 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-ss8nf"] Oct 01 09:45:42 crc kubenswrapper[4787]: E1001 09:45:42.568340 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c159d29b-be19-4d0f-980e-a12c2ed641f2" containerName="collect-profiles" Oct 01 09:45:42 crc kubenswrapper[4787]: I1001 09:45:42.568351 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c159d29b-be19-4d0f-980e-a12c2ed641f2" containerName="collect-profiles" Oct 01 09:45:42 crc kubenswrapper[4787]: I1001 09:45:42.568443 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="c159d29b-be19-4d0f-980e-a12c2ed641f2" containerName="collect-profiles" Oct 01 09:45:42 crc kubenswrapper[4787]: I1001 09:45:42.568805 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-ss8nf" Oct 01 09:45:42 crc kubenswrapper[4787]: I1001 09:45:42.571508 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 01 09:45:42 crc kubenswrapper[4787]: I1001 09:45:42.571698 4787 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-cqxcw" Oct 01 09:45:42 crc kubenswrapper[4787]: I1001 09:45:42.572131 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 01 09:45:42 crc kubenswrapper[4787]: I1001 09:45:42.578159 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-ss8nf"] Oct 01 09:45:42 crc kubenswrapper[4787]: I1001 09:45:42.583504 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-w2pvt"] Oct 01 09:45:42 crc kubenswrapper[4787]: I1001 09:45:42.589105 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-w2pvt" Oct 01 09:45:42 crc kubenswrapper[4787]: I1001 09:45:42.592857 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-w2pvt"] Oct 01 09:45:42 crc kubenswrapper[4787]: I1001 09:45:42.592953 4787 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-5lqsz" Oct 01 09:45:42 crc kubenswrapper[4787]: I1001 09:45:42.611456 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qq7zn\" (UniqueName: \"kubernetes.io/projected/081d3606-6bc1-4f9b-8f3b-6ee00ed9d164-kube-api-access-qq7zn\") pod \"cert-manager-cainjector-7f985d654d-ss8nf\" (UID: \"081d3606-6bc1-4f9b-8f3b-6ee00ed9d164\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-ss8nf" Oct 01 09:45:42 crc kubenswrapper[4787]: I1001 09:45:42.611586 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qqbc\" (UniqueName: \"kubernetes.io/projected/b1f30c7b-06ed-4ccf-a3d0-8df65a3a8c67-kube-api-access-6qqbc\") pod \"cert-manager-5b446d88c5-w2pvt\" (UID: \"b1f30c7b-06ed-4ccf-a3d0-8df65a3a8c67\") " pod="cert-manager/cert-manager-5b446d88c5-w2pvt" Oct 01 09:45:42 crc kubenswrapper[4787]: I1001 09:45:42.614617 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-xgrnp"] Oct 01 09:45:42 crc kubenswrapper[4787]: I1001 09:45:42.615825 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-xgrnp" Oct 01 09:45:42 crc kubenswrapper[4787]: I1001 09:45:42.618029 4787 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-hw24g" Oct 01 09:45:42 crc kubenswrapper[4787]: I1001 09:45:42.621154 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-xgrnp"] Oct 01 09:45:42 crc kubenswrapper[4787]: I1001 09:45:42.712641 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qqbc\" (UniqueName: \"kubernetes.io/projected/b1f30c7b-06ed-4ccf-a3d0-8df65a3a8c67-kube-api-access-6qqbc\") pod \"cert-manager-5b446d88c5-w2pvt\" (UID: \"b1f30c7b-06ed-4ccf-a3d0-8df65a3a8c67\") " pod="cert-manager/cert-manager-5b446d88c5-w2pvt" Oct 01 09:45:42 crc kubenswrapper[4787]: I1001 09:45:42.712728 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vh2nt\" (UniqueName: \"kubernetes.io/projected/426ff9a0-bcf0-486e-bd80-2d4de51be0cf-kube-api-access-vh2nt\") pod \"cert-manager-webhook-5655c58dd6-xgrnp\" (UID: \"426ff9a0-bcf0-486e-bd80-2d4de51be0cf\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-xgrnp" Oct 01 09:45:42 crc kubenswrapper[4787]: I1001 09:45:42.712805 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qq7zn\" (UniqueName: \"kubernetes.io/projected/081d3606-6bc1-4f9b-8f3b-6ee00ed9d164-kube-api-access-qq7zn\") pod \"cert-manager-cainjector-7f985d654d-ss8nf\" (UID: \"081d3606-6bc1-4f9b-8f3b-6ee00ed9d164\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-ss8nf" Oct 01 09:45:42 crc kubenswrapper[4787]: I1001 09:45:42.729944 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qqbc\" (UniqueName: \"kubernetes.io/projected/b1f30c7b-06ed-4ccf-a3d0-8df65a3a8c67-kube-api-access-6qqbc\") pod \"cert-manager-5b446d88c5-w2pvt\" (UID: \"b1f30c7b-06ed-4ccf-a3d0-8df65a3a8c67\") " pod="cert-manager/cert-manager-5b446d88c5-w2pvt" Oct 01 09:45:42 crc kubenswrapper[4787]: I1001 09:45:42.732190 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qq7zn\" (UniqueName: \"kubernetes.io/projected/081d3606-6bc1-4f9b-8f3b-6ee00ed9d164-kube-api-access-qq7zn\") pod \"cert-manager-cainjector-7f985d654d-ss8nf\" (UID: \"081d3606-6bc1-4f9b-8f3b-6ee00ed9d164\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-ss8nf" Oct 01 09:45:42 crc kubenswrapper[4787]: I1001 09:45:42.813508 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vh2nt\" (UniqueName: \"kubernetes.io/projected/426ff9a0-bcf0-486e-bd80-2d4de51be0cf-kube-api-access-vh2nt\") pod \"cert-manager-webhook-5655c58dd6-xgrnp\" (UID: \"426ff9a0-bcf0-486e-bd80-2d4de51be0cf\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-xgrnp" Oct 01 09:45:42 crc kubenswrapper[4787]: I1001 09:45:42.833774 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vh2nt\" (UniqueName: \"kubernetes.io/projected/426ff9a0-bcf0-486e-bd80-2d4de51be0cf-kube-api-access-vh2nt\") pod \"cert-manager-webhook-5655c58dd6-xgrnp\" (UID: \"426ff9a0-bcf0-486e-bd80-2d4de51be0cf\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-xgrnp" Oct 01 09:45:42 crc kubenswrapper[4787]: I1001 09:45:42.887501 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-ss8nf" Oct 01 09:45:42 crc kubenswrapper[4787]: I1001 09:45:42.906663 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-w2pvt" Oct 01 09:45:42 crc kubenswrapper[4787]: I1001 09:45:42.930825 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-xgrnp" Oct 01 09:45:43 crc kubenswrapper[4787]: I1001 09:45:43.117414 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-ss8nf"] Oct 01 09:45:43 crc kubenswrapper[4787]: I1001 09:45:43.129129 4787 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 09:45:43 crc kubenswrapper[4787]: I1001 09:45:43.173325 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-w2pvt"] Oct 01 09:45:43 crc kubenswrapper[4787]: I1001 09:45:43.208275 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-ss8nf" event={"ID":"081d3606-6bc1-4f9b-8f3b-6ee00ed9d164","Type":"ContainerStarted","Data":"fd7a056ed4234570e2daf76d29e583217f003ca9b447b62e4474fc87b87355d1"} Oct 01 09:45:43 crc kubenswrapper[4787]: I1001 09:45:43.213660 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-w2pvt" event={"ID":"b1f30c7b-06ed-4ccf-a3d0-8df65a3a8c67","Type":"ContainerStarted","Data":"23791f3c2c9adc9ab75f7bf0276a850df2a823114b7f0cf77ab2ce8f41175286"} Oct 01 09:45:43 crc kubenswrapper[4787]: I1001 09:45:43.215455 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-xgrnp"] Oct 01 09:45:44 crc kubenswrapper[4787]: I1001 09:45:44.222725 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-xgrnp" event={"ID":"426ff9a0-bcf0-486e-bd80-2d4de51be0cf","Type":"ContainerStarted","Data":"22f04454bdf164afcbbb96ecf17ad95e99d2471399024b53fe55d837849e66d6"} Oct 01 09:45:46 crc kubenswrapper[4787]: I1001 09:45:46.240299 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-xgrnp" event={"ID":"426ff9a0-bcf0-486e-bd80-2d4de51be0cf","Type":"ContainerStarted","Data":"89bf441301b3917d12a762e665f2a421c95770fbf3b0e2989339760b175b3a45"} Oct 01 09:45:46 crc kubenswrapper[4787]: I1001 09:45:46.240703 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-xgrnp" Oct 01 09:45:46 crc kubenswrapper[4787]: I1001 09:45:46.255812 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-xgrnp" podStartSLOduration=1.425417816 podStartE2EDuration="4.255790907s" podCreationTimestamp="2025-10-01 09:45:42 +0000 UTC" firstStartedPulling="2025-10-01 09:45:43.221336977 +0000 UTC m=+575.336481134" lastFinishedPulling="2025-10-01 09:45:46.051710068 +0000 UTC m=+578.166854225" observedRunningTime="2025-10-01 09:45:46.25321686 +0000 UTC m=+578.368361037" watchObservedRunningTime="2025-10-01 09:45:46.255790907 +0000 UTC m=+578.370935074" Oct 01 09:45:47 crc kubenswrapper[4787]: I1001 09:45:47.247136 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-ss8nf" event={"ID":"081d3606-6bc1-4f9b-8f3b-6ee00ed9d164","Type":"ContainerStarted","Data":"7340dcdb8647a62f47ce1105dd15d3b77822638098cb74d2570f58545feaf9b5"} Oct 01 09:45:47 crc kubenswrapper[4787]: I1001 09:45:47.249012 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-w2pvt" event={"ID":"b1f30c7b-06ed-4ccf-a3d0-8df65a3a8c67","Type":"ContainerStarted","Data":"6bd8aa396aae092d8c3d9fe5aa13291546c9561d73db736f19ebed84280b7ecd"} Oct 01 09:45:47 crc kubenswrapper[4787]: I1001 09:45:47.268902 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-ss8nf" podStartSLOduration=2.399054066 podStartE2EDuration="5.268879901s" podCreationTimestamp="2025-10-01 09:45:42 +0000 UTC" firstStartedPulling="2025-10-01 09:45:43.128853096 +0000 UTC m=+575.243997253" lastFinishedPulling="2025-10-01 09:45:45.998678931 +0000 UTC m=+578.113823088" observedRunningTime="2025-10-01 09:45:47.268180254 +0000 UTC m=+579.383324431" watchObservedRunningTime="2025-10-01 09:45:47.268879901 +0000 UTC m=+579.384024048" Oct 01 09:45:47 crc kubenswrapper[4787]: I1001 09:45:47.283031 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-w2pvt" podStartSLOduration=2.350942906 podStartE2EDuration="5.282999758s" podCreationTimestamp="2025-10-01 09:45:42 +0000 UTC" firstStartedPulling="2025-10-01 09:45:43.183758101 +0000 UTC m=+575.298902258" lastFinishedPulling="2025-10-01 09:45:46.115814953 +0000 UTC m=+578.230959110" observedRunningTime="2025-10-01 09:45:47.281638052 +0000 UTC m=+579.396782219" watchObservedRunningTime="2025-10-01 09:45:47.282999758 +0000 UTC m=+579.398143915" Oct 01 09:45:52 crc kubenswrapper[4787]: I1001 09:45:52.934821 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-xgrnp" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.326530 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-dmsxm"] Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.326983 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="ovn-controller" containerID="cri-o://394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694" gracePeriod=30 Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.327163 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="northd" containerID="cri-o://834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e" gracePeriod=30 Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.327225 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274" gracePeriod=30 Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.327172 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="nbdb" containerID="cri-o://e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4" gracePeriod=30 Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.327269 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="ovn-acl-logging" containerID="cri-o://b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a" gracePeriod=30 Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.327399 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="kube-rbac-proxy-node" containerID="cri-o://7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f" gracePeriod=30 Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.327509 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="sbdb" containerID="cri-o://85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68" gracePeriod=30 Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.380178 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="ovnkube-controller" containerID="cri-o://81f37f28240c3a86515abe10685e1a875a8cafa2df45d92b131677f3b1eacdac" gracePeriod=30 Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.671208 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dmsxm_eadf2545-886e-4642-b909-704c6cd4134e/ovnkube-controller/3.log" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.674272 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dmsxm_eadf2545-886e-4642-b909-704c6cd4134e/ovn-acl-logging/0.log" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.675100 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dmsxm_eadf2545-886e-4642-b909-704c6cd4134e/ovn-controller/0.log" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.675608 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.731660 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-vv8p9"] Oct 01 09:45:53 crc kubenswrapper[4787]: E1001 09:45:53.731883 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="ovnkube-controller" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.731901 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="ovnkube-controller" Oct 01 09:45:53 crc kubenswrapper[4787]: E1001 09:45:53.731911 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="kube-rbac-proxy-ovn-metrics" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.731918 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="kube-rbac-proxy-ovn-metrics" Oct 01 09:45:53 crc kubenswrapper[4787]: E1001 09:45:53.731932 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="ovn-controller" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.731939 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="ovn-controller" Oct 01 09:45:53 crc kubenswrapper[4787]: E1001 09:45:53.731947 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="nbdb" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.731954 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="nbdb" Oct 01 09:45:53 crc kubenswrapper[4787]: E1001 09:45:53.731965 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="kubecfg-setup" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.731972 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="kubecfg-setup" Oct 01 09:45:53 crc kubenswrapper[4787]: E1001 09:45:53.731983 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="sbdb" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.731990 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="sbdb" Oct 01 09:45:53 crc kubenswrapper[4787]: E1001 09:45:53.731998 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="kube-rbac-proxy-node" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.732005 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="kube-rbac-proxy-node" Oct 01 09:45:53 crc kubenswrapper[4787]: E1001 09:45:53.732016 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="ovnkube-controller" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.732024 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="ovnkube-controller" Oct 01 09:45:53 crc kubenswrapper[4787]: E1001 09:45:53.732031 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="ovnkube-controller" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.732040 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="ovnkube-controller" Oct 01 09:45:53 crc kubenswrapper[4787]: E1001 09:45:53.732055 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="ovn-acl-logging" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.732063 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="ovn-acl-logging" Oct 01 09:45:53 crc kubenswrapper[4787]: E1001 09:45:53.732093 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="northd" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.732101 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="northd" Oct 01 09:45:53 crc kubenswrapper[4787]: E1001 09:45:53.732114 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="ovnkube-controller" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.732122 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="ovnkube-controller" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.732242 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="ovnkube-controller" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.732256 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="kube-rbac-proxy-node" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.732266 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="kube-rbac-proxy-ovn-metrics" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.732275 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="ovn-controller" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.732285 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="ovn-acl-logging" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.732297 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="sbdb" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.732307 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="ovnkube-controller" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.732316 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="nbdb" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.732325 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="northd" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.732334 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="ovnkube-controller" Oct 01 09:45:53 crc kubenswrapper[4787]: E1001 09:45:53.732440 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="ovnkube-controller" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.732449 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="ovnkube-controller" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.732572 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="ovnkube-controller" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.732582 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="eadf2545-886e-4642-b909-704c6cd4134e" containerName="ovnkube-controller" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.734108 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.811924 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"eadf2545-886e-4642-b909-704c6cd4134e\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.812342 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-run-netns\") pod \"eadf2545-886e-4642-b909-704c6cd4134e\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.812059 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "eadf2545-886e-4642-b909-704c6cd4134e" (UID: "eadf2545-886e-4642-b909-704c6cd4134e"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.812391 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/eadf2545-886e-4642-b909-704c6cd4134e-ovnkube-config\") pod \"eadf2545-886e-4642-b909-704c6cd4134e\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.812410 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "eadf2545-886e-4642-b909-704c6cd4134e" (UID: "eadf2545-886e-4642-b909-704c6cd4134e"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.812430 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8kgg\" (UniqueName: \"kubernetes.io/projected/eadf2545-886e-4642-b909-704c6cd4134e-kube-api-access-z8kgg\") pod \"eadf2545-886e-4642-b909-704c6cd4134e\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.812460 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-cni-netd\") pod \"eadf2545-886e-4642-b909-704c6cd4134e\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.812483 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-run-ovn-kubernetes\") pod \"eadf2545-886e-4642-b909-704c6cd4134e\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.812536 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-log-socket\") pod \"eadf2545-886e-4642-b909-704c6cd4134e\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.812570 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-kubelet\") pod \"eadf2545-886e-4642-b909-704c6cd4134e\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.812595 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "eadf2545-886e-4642-b909-704c6cd4134e" (UID: "eadf2545-886e-4642-b909-704c6cd4134e"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.812634 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-log-socket" (OuterVolumeSpecName: "log-socket") pod "eadf2545-886e-4642-b909-704c6cd4134e" (UID: "eadf2545-886e-4642-b909-704c6cd4134e"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.812652 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-run-ovn\") pod \"eadf2545-886e-4642-b909-704c6cd4134e\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.812677 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-var-lib-openvswitch\") pod \"eadf2545-886e-4642-b909-704c6cd4134e\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.812689 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "eadf2545-886e-4642-b909-704c6cd4134e" (UID: "eadf2545-886e-4642-b909-704c6cd4134e"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.812702 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/eadf2545-886e-4642-b909-704c6cd4134e-env-overrides\") pod \"eadf2545-886e-4642-b909-704c6cd4134e\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.812733 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "eadf2545-886e-4642-b909-704c6cd4134e" (UID: "eadf2545-886e-4642-b909-704c6cd4134e"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.812737 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-slash\") pod \"eadf2545-886e-4642-b909-704c6cd4134e\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.812739 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "eadf2545-886e-4642-b909-704c6cd4134e" (UID: "eadf2545-886e-4642-b909-704c6cd4134e"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.812755 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "eadf2545-886e-4642-b909-704c6cd4134e" (UID: "eadf2545-886e-4642-b909-704c6cd4134e"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.812774 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-slash" (OuterVolumeSpecName: "host-slash") pod "eadf2545-886e-4642-b909-704c6cd4134e" (UID: "eadf2545-886e-4642-b909-704c6cd4134e"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.812793 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/eadf2545-886e-4642-b909-704c6cd4134e-ovnkube-script-lib\") pod \"eadf2545-886e-4642-b909-704c6cd4134e\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.812853 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eadf2545-886e-4642-b909-704c6cd4134e-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "eadf2545-886e-4642-b909-704c6cd4134e" (UID: "eadf2545-886e-4642-b909-704c6cd4134e"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.812857 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-run-openvswitch\") pod \"eadf2545-886e-4642-b909-704c6cd4134e\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.812885 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-systemd-units\") pod \"eadf2545-886e-4642-b909-704c6cd4134e\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.812890 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "eadf2545-886e-4642-b909-704c6cd4134e" (UID: "eadf2545-886e-4642-b909-704c6cd4134e"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.812912 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-cni-bin\") pod \"eadf2545-886e-4642-b909-704c6cd4134e\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.812930 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "eadf2545-886e-4642-b909-704c6cd4134e" (UID: "eadf2545-886e-4642-b909-704c6cd4134e"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.812955 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/eadf2545-886e-4642-b909-704c6cd4134e-ovn-node-metrics-cert\") pod \"eadf2545-886e-4642-b909-704c6cd4134e\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.812981 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "eadf2545-886e-4642-b909-704c6cd4134e" (UID: "eadf2545-886e-4642-b909-704c6cd4134e"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.812997 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-run-systemd\") pod \"eadf2545-886e-4642-b909-704c6cd4134e\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.813030 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-node-log\") pod \"eadf2545-886e-4642-b909-704c6cd4134e\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.813051 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-etc-openvswitch\") pod \"eadf2545-886e-4642-b909-704c6cd4134e\" (UID: \"eadf2545-886e-4642-b909-704c6cd4134e\") " Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.813172 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-node-log" (OuterVolumeSpecName: "node-log") pod "eadf2545-886e-4642-b909-704c6cd4134e" (UID: "eadf2545-886e-4642-b909-704c6cd4134e"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.813204 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "eadf2545-886e-4642-b909-704c6cd4134e" (UID: "eadf2545-886e-4642-b909-704c6cd4134e"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.813229 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eadf2545-886e-4642-b909-704c6cd4134e-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "eadf2545-886e-4642-b909-704c6cd4134e" (UID: "eadf2545-886e-4642-b909-704c6cd4134e"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.813255 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-host-run-netns\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.813286 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-host-cni-bin\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.813280 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eadf2545-886e-4642-b909-704c6cd4134e-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "eadf2545-886e-4642-b909-704c6cd4134e" (UID: "eadf2545-886e-4642-b909-704c6cd4134e"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.813320 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-host-cni-netd\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.813344 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.813396 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-systemd-units\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.813474 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-etc-openvswitch\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.813513 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/29fc1929-bc7b-4340-afd7-f801f43909a1-env-overrides\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.813538 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-host-slash\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.813668 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-run-ovn\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.813697 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-run-systemd\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.813724 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-node-log\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.813753 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-log-socket\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.813786 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-host-kubelet\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.813811 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-run-openvswitch\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.813834 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-host-run-ovn-kubernetes\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.813863 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98279\" (UniqueName: \"kubernetes.io/projected/29fc1929-bc7b-4340-afd7-f801f43909a1-kube-api-access-98279\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.813910 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/29fc1929-bc7b-4340-afd7-f801f43909a1-ovnkube-script-lib\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.813934 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-var-lib-openvswitch\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.813959 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/29fc1929-bc7b-4340-afd7-f801f43909a1-ovnkube-config\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.813981 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/29fc1929-bc7b-4340-afd7-f801f43909a1-ovn-node-metrics-cert\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.814065 4787 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.814101 4787 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.814114 4787 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.814127 4787 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/eadf2545-886e-4642-b909-704c6cd4134e-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.814139 4787 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-slash\") on node \"crc\" DevicePath \"\"" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.814151 4787 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/eadf2545-886e-4642-b909-704c6cd4134e-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.814163 4787 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.814174 4787 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.814186 4787 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.814199 4787 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-node-log\") on node \"crc\" DevicePath \"\"" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.814210 4787 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.814223 4787 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.814236 4787 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.814249 4787 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/eadf2545-886e-4642-b909-704c6cd4134e-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.814261 4787 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.814274 4787 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.814286 4787 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-log-socket\") on node \"crc\" DevicePath \"\"" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.820218 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eadf2545-886e-4642-b909-704c6cd4134e-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "eadf2545-886e-4642-b909-704c6cd4134e" (UID: "eadf2545-886e-4642-b909-704c6cd4134e"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.820984 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eadf2545-886e-4642-b909-704c6cd4134e-kube-api-access-z8kgg" (OuterVolumeSpecName: "kube-api-access-z8kgg") pod "eadf2545-886e-4642-b909-704c6cd4134e" (UID: "eadf2545-886e-4642-b909-704c6cd4134e"). InnerVolumeSpecName "kube-api-access-z8kgg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.827643 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "eadf2545-886e-4642-b909-704c6cd4134e" (UID: "eadf2545-886e-4642-b909-704c6cd4134e"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.915876 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/29fc1929-bc7b-4340-afd7-f801f43909a1-ovnkube-script-lib\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.915932 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-var-lib-openvswitch\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.915961 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/29fc1929-bc7b-4340-afd7-f801f43909a1-ovnkube-config\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.915985 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/29fc1929-bc7b-4340-afd7-f801f43909a1-ovn-node-metrics-cert\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.916029 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-host-run-netns\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.916051 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-host-cni-bin\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.916092 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-host-cni-netd\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.916116 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.916142 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-systemd-units\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.916180 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-etc-openvswitch\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.916201 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/29fc1929-bc7b-4340-afd7-f801f43909a1-env-overrides\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.916221 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-host-slash\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.916270 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-run-ovn\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.916291 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-run-systemd\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.916315 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-node-log\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.916335 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-log-socket\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.916360 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-host-kubelet\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.916379 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-run-openvswitch\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.916402 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-host-run-ovn-kubernetes\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.916428 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98279\" (UniqueName: \"kubernetes.io/projected/29fc1929-bc7b-4340-afd7-f801f43909a1-kube-api-access-98279\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.916511 4787 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/eadf2545-886e-4642-b909-704c6cd4134e-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.916527 4787 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/eadf2545-886e-4642-b909-704c6cd4134e-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.916541 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8kgg\" (UniqueName: \"kubernetes.io/projected/eadf2545-886e-4642-b909-704c6cd4134e-kube-api-access-z8kgg\") on node \"crc\" DevicePath \"\"" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.917297 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-etc-openvswitch\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.917409 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-host-run-netns\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.917445 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-host-cni-netd\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.917406 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-node-log\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.917489 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-host-slash\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.917408 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-host-cni-bin\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.917577 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-run-systemd\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.917599 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-var-lib-openvswitch\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.917628 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.917640 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-run-ovn\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.917674 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/29fc1929-bc7b-4340-afd7-f801f43909a1-ovnkube-script-lib\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.917690 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-log-socket\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.917674 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-systemd-units\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.917736 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-host-kubelet\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.917766 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-run-openvswitch\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.917793 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/29fc1929-bc7b-4340-afd7-f801f43909a1-host-run-ovn-kubernetes\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.918397 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/29fc1929-bc7b-4340-afd7-f801f43909a1-env-overrides\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.918715 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/29fc1929-bc7b-4340-afd7-f801f43909a1-ovnkube-config\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.921243 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/29fc1929-bc7b-4340-afd7-f801f43909a1-ovn-node-metrics-cert\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:53 crc kubenswrapper[4787]: I1001 09:45:53.939421 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98279\" (UniqueName: \"kubernetes.io/projected/29fc1929-bc7b-4340-afd7-f801f43909a1-kube-api-access-98279\") pod \"ovnkube-node-vv8p9\" (UID: \"29fc1929-bc7b-4340-afd7-f801f43909a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.052856 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.329907 4787 generic.go:334] "Generic (PLEG): container finished" podID="29fc1929-bc7b-4340-afd7-f801f43909a1" containerID="c9fdd8e29ae66a1fd89f013013b06e1a3d10ae96d9beaad7be42e77350bb37c8" exitCode=0 Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.329987 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" event={"ID":"29fc1929-bc7b-4340-afd7-f801f43909a1","Type":"ContainerDied","Data":"c9fdd8e29ae66a1fd89f013013b06e1a3d10ae96d9beaad7be42e77350bb37c8"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.330037 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" event={"ID":"29fc1929-bc7b-4340-afd7-f801f43909a1","Type":"ContainerStarted","Data":"ddc60c55102bc6d9504c47d6386042129c544bb72ff8aa635c7ff61acf11bbeb"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.332469 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z8tjz_53f5ffd2-7649-4c9e-929f-a68187409644/kube-multus/2.log" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.334673 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z8tjz_53f5ffd2-7649-4c9e-929f-a68187409644/kube-multus/1.log" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.334718 4787 generic.go:334] "Generic (PLEG): container finished" podID="53f5ffd2-7649-4c9e-929f-a68187409644" containerID="8ccabd0db928904435aacf77b37e997bf489a08e4cc113b3438d986e0c7f0ed2" exitCode=2 Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.334745 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-z8tjz" event={"ID":"53f5ffd2-7649-4c9e-929f-a68187409644","Type":"ContainerDied","Data":"8ccabd0db928904435aacf77b37e997bf489a08e4cc113b3438d986e0c7f0ed2"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.334789 4787 scope.go:117] "RemoveContainer" containerID="8c830596f83deb4fa998fa9af881c7eb1f3673588f444edc9cbf895c19b8248c" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.335310 4787 scope.go:117] "RemoveContainer" containerID="8ccabd0db928904435aacf77b37e997bf489a08e4cc113b3438d986e0c7f0ed2" Oct 01 09:45:54 crc kubenswrapper[4787]: E1001 09:45:54.335622 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-z8tjz_openshift-multus(53f5ffd2-7649-4c9e-929f-a68187409644)\"" pod="openshift-multus/multus-z8tjz" podUID="53f5ffd2-7649-4c9e-929f-a68187409644" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.337577 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dmsxm_eadf2545-886e-4642-b909-704c6cd4134e/ovnkube-controller/3.log" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.340048 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dmsxm_eadf2545-886e-4642-b909-704c6cd4134e/ovn-acl-logging/0.log" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.340787 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dmsxm_eadf2545-886e-4642-b909-704c6cd4134e/ovn-controller/0.log" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341244 4787 generic.go:334] "Generic (PLEG): container finished" podID="eadf2545-886e-4642-b909-704c6cd4134e" containerID="81f37f28240c3a86515abe10685e1a875a8cafa2df45d92b131677f3b1eacdac" exitCode=0 Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341272 4787 generic.go:334] "Generic (PLEG): container finished" podID="eadf2545-886e-4642-b909-704c6cd4134e" containerID="85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68" exitCode=0 Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341285 4787 generic.go:334] "Generic (PLEG): container finished" podID="eadf2545-886e-4642-b909-704c6cd4134e" containerID="e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4" exitCode=0 Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341296 4787 generic.go:334] "Generic (PLEG): container finished" podID="eadf2545-886e-4642-b909-704c6cd4134e" containerID="834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e" exitCode=0 Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341306 4787 generic.go:334] "Generic (PLEG): container finished" podID="eadf2545-886e-4642-b909-704c6cd4134e" containerID="39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274" exitCode=0 Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341318 4787 generic.go:334] "Generic (PLEG): container finished" podID="eadf2545-886e-4642-b909-704c6cd4134e" containerID="7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f" exitCode=0 Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341329 4787 generic.go:334] "Generic (PLEG): container finished" podID="eadf2545-886e-4642-b909-704c6cd4134e" containerID="b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a" exitCode=143 Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341339 4787 generic.go:334] "Generic (PLEG): container finished" podID="eadf2545-886e-4642-b909-704c6cd4134e" containerID="394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694" exitCode=143 Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341312 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" event={"ID":"eadf2545-886e-4642-b909-704c6cd4134e","Type":"ContainerDied","Data":"81f37f28240c3a86515abe10685e1a875a8cafa2df45d92b131677f3b1eacdac"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341392 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341409 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" event={"ID":"eadf2545-886e-4642-b909-704c6cd4134e","Type":"ContainerDied","Data":"85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341438 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" event={"ID":"eadf2545-886e-4642-b909-704c6cd4134e","Type":"ContainerDied","Data":"e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341460 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" event={"ID":"eadf2545-886e-4642-b909-704c6cd4134e","Type":"ContainerDied","Data":"834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341479 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" event={"ID":"eadf2545-886e-4642-b909-704c6cd4134e","Type":"ContainerDied","Data":"39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341500 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" event={"ID":"eadf2545-886e-4642-b909-704c6cd4134e","Type":"ContainerDied","Data":"7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341518 4787 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"81f37f28240c3a86515abe10685e1a875a8cafa2df45d92b131677f3b1eacdac"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341536 4787 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341548 4787 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341559 4787 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341594 4787 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341606 4787 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341617 4787 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341629 4787 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341642 4787 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341653 4787 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341669 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" event={"ID":"eadf2545-886e-4642-b909-704c6cd4134e","Type":"ContainerDied","Data":"b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341687 4787 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"81f37f28240c3a86515abe10685e1a875a8cafa2df45d92b131677f3b1eacdac"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341701 4787 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341714 4787 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341725 4787 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341737 4787 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341748 4787 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341778 4787 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341790 4787 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341801 4787 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341813 4787 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341828 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" event={"ID":"eadf2545-886e-4642-b909-704c6cd4134e","Type":"ContainerDied","Data":"394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341845 4787 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"81f37f28240c3a86515abe10685e1a875a8cafa2df45d92b131677f3b1eacdac"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341858 4787 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341869 4787 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341880 4787 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341892 4787 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341902 4787 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341915 4787 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341926 4787 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341940 4787 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341951 4787 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341965 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dmsxm" event={"ID":"eadf2545-886e-4642-b909-704c6cd4134e","Type":"ContainerDied","Data":"2f1d4c83eb343b113400d9faa34fc103f326c7ff17e9312a4aa55d85f97106eb"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341983 4787 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"81f37f28240c3a86515abe10685e1a875a8cafa2df45d92b131677f3b1eacdac"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.341999 4787 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.342010 4787 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.342020 4787 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.342031 4787 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.342043 4787 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.342053 4787 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.342064 4787 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.342100 4787 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.342111 4787 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c"} Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.374868 4787 scope.go:117] "RemoveContainer" containerID="81f37f28240c3a86515abe10685e1a875a8cafa2df45d92b131677f3b1eacdac" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.404810 4787 scope.go:117] "RemoveContainer" containerID="c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.413445 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-dmsxm"] Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.418214 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-dmsxm"] Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.424683 4787 scope.go:117] "RemoveContainer" containerID="85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.450701 4787 scope.go:117] "RemoveContainer" containerID="e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.485571 4787 scope.go:117] "RemoveContainer" containerID="834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.500905 4787 scope.go:117] "RemoveContainer" containerID="39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.515928 4787 scope.go:117] "RemoveContainer" containerID="7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.531717 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eadf2545-886e-4642-b909-704c6cd4134e" path="/var/lib/kubelet/pods/eadf2545-886e-4642-b909-704c6cd4134e/volumes" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.532268 4787 scope.go:117] "RemoveContainer" containerID="b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.547155 4787 scope.go:117] "RemoveContainer" containerID="394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.575512 4787 scope.go:117] "RemoveContainer" containerID="a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.593256 4787 scope.go:117] "RemoveContainer" containerID="81f37f28240c3a86515abe10685e1a875a8cafa2df45d92b131677f3b1eacdac" Oct 01 09:45:54 crc kubenswrapper[4787]: E1001 09:45:54.593667 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81f37f28240c3a86515abe10685e1a875a8cafa2df45d92b131677f3b1eacdac\": container with ID starting with 81f37f28240c3a86515abe10685e1a875a8cafa2df45d92b131677f3b1eacdac not found: ID does not exist" containerID="81f37f28240c3a86515abe10685e1a875a8cafa2df45d92b131677f3b1eacdac" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.593709 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81f37f28240c3a86515abe10685e1a875a8cafa2df45d92b131677f3b1eacdac"} err="failed to get container status \"81f37f28240c3a86515abe10685e1a875a8cafa2df45d92b131677f3b1eacdac\": rpc error: code = NotFound desc = could not find container \"81f37f28240c3a86515abe10685e1a875a8cafa2df45d92b131677f3b1eacdac\": container with ID starting with 81f37f28240c3a86515abe10685e1a875a8cafa2df45d92b131677f3b1eacdac not found: ID does not exist" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.593735 4787 scope.go:117] "RemoveContainer" containerID="c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122" Oct 01 09:45:54 crc kubenswrapper[4787]: E1001 09:45:54.594152 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122\": container with ID starting with c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122 not found: ID does not exist" containerID="c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.594178 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122"} err="failed to get container status \"c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122\": rpc error: code = NotFound desc = could not find container \"c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122\": container with ID starting with c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122 not found: ID does not exist" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.594197 4787 scope.go:117] "RemoveContainer" containerID="85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68" Oct 01 09:45:54 crc kubenswrapper[4787]: E1001 09:45:54.594513 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68\": container with ID starting with 85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68 not found: ID does not exist" containerID="85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.594564 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68"} err="failed to get container status \"85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68\": rpc error: code = NotFound desc = could not find container \"85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68\": container with ID starting with 85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68 not found: ID does not exist" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.594603 4787 scope.go:117] "RemoveContainer" containerID="e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4" Oct 01 09:45:54 crc kubenswrapper[4787]: E1001 09:45:54.594922 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4\": container with ID starting with e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4 not found: ID does not exist" containerID="e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.594950 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4"} err="failed to get container status \"e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4\": rpc error: code = NotFound desc = could not find container \"e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4\": container with ID starting with e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4 not found: ID does not exist" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.594966 4787 scope.go:117] "RemoveContainer" containerID="834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e" Oct 01 09:45:54 crc kubenswrapper[4787]: E1001 09:45:54.595648 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e\": container with ID starting with 834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e not found: ID does not exist" containerID="834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.595697 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e"} err="failed to get container status \"834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e\": rpc error: code = NotFound desc = could not find container \"834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e\": container with ID starting with 834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e not found: ID does not exist" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.595730 4787 scope.go:117] "RemoveContainer" containerID="39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274" Oct 01 09:45:54 crc kubenswrapper[4787]: E1001 09:45:54.596672 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274\": container with ID starting with 39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274 not found: ID does not exist" containerID="39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.596714 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274"} err="failed to get container status \"39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274\": rpc error: code = NotFound desc = could not find container \"39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274\": container with ID starting with 39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274 not found: ID does not exist" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.596731 4787 scope.go:117] "RemoveContainer" containerID="7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f" Oct 01 09:45:54 crc kubenswrapper[4787]: E1001 09:45:54.597010 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f\": container with ID starting with 7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f not found: ID does not exist" containerID="7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.597053 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f"} err="failed to get container status \"7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f\": rpc error: code = NotFound desc = could not find container \"7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f\": container with ID starting with 7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f not found: ID does not exist" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.597104 4787 scope.go:117] "RemoveContainer" containerID="b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a" Oct 01 09:45:54 crc kubenswrapper[4787]: E1001 09:45:54.597385 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a\": container with ID starting with b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a not found: ID does not exist" containerID="b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.597418 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a"} err="failed to get container status \"b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a\": rpc error: code = NotFound desc = could not find container \"b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a\": container with ID starting with b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a not found: ID does not exist" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.597442 4787 scope.go:117] "RemoveContainer" containerID="394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694" Oct 01 09:45:54 crc kubenswrapper[4787]: E1001 09:45:54.597704 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694\": container with ID starting with 394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694 not found: ID does not exist" containerID="394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.597737 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694"} err="failed to get container status \"394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694\": rpc error: code = NotFound desc = could not find container \"394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694\": container with ID starting with 394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694 not found: ID does not exist" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.597760 4787 scope.go:117] "RemoveContainer" containerID="a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c" Oct 01 09:45:54 crc kubenswrapper[4787]: E1001 09:45:54.598031 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\": container with ID starting with a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c not found: ID does not exist" containerID="a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.598070 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c"} err="failed to get container status \"a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\": rpc error: code = NotFound desc = could not find container \"a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\": container with ID starting with a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c not found: ID does not exist" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.598121 4787 scope.go:117] "RemoveContainer" containerID="81f37f28240c3a86515abe10685e1a875a8cafa2df45d92b131677f3b1eacdac" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.598367 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81f37f28240c3a86515abe10685e1a875a8cafa2df45d92b131677f3b1eacdac"} err="failed to get container status \"81f37f28240c3a86515abe10685e1a875a8cafa2df45d92b131677f3b1eacdac\": rpc error: code = NotFound desc = could not find container \"81f37f28240c3a86515abe10685e1a875a8cafa2df45d92b131677f3b1eacdac\": container with ID starting with 81f37f28240c3a86515abe10685e1a875a8cafa2df45d92b131677f3b1eacdac not found: ID does not exist" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.598394 4787 scope.go:117] "RemoveContainer" containerID="c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.598609 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122"} err="failed to get container status \"c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122\": rpc error: code = NotFound desc = could not find container \"c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122\": container with ID starting with c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122 not found: ID does not exist" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.598633 4787 scope.go:117] "RemoveContainer" containerID="85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.598867 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68"} err="failed to get container status \"85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68\": rpc error: code = NotFound desc = could not find container \"85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68\": container with ID starting with 85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68 not found: ID does not exist" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.598888 4787 scope.go:117] "RemoveContainer" containerID="e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.599172 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4"} err="failed to get container status \"e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4\": rpc error: code = NotFound desc = could not find container \"e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4\": container with ID starting with e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4 not found: ID does not exist" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.599202 4787 scope.go:117] "RemoveContainer" containerID="834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.599484 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e"} err="failed to get container status \"834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e\": rpc error: code = NotFound desc = could not find container \"834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e\": container with ID starting with 834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e not found: ID does not exist" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.599520 4787 scope.go:117] "RemoveContainer" containerID="39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.599781 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274"} err="failed to get container status \"39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274\": rpc error: code = NotFound desc = could not find container \"39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274\": container with ID starting with 39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274 not found: ID does not exist" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.599816 4787 scope.go:117] "RemoveContainer" containerID="7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.600160 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f"} err="failed to get container status \"7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f\": rpc error: code = NotFound desc = could not find container \"7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f\": container with ID starting with 7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f not found: ID does not exist" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.600197 4787 scope.go:117] "RemoveContainer" containerID="b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.600442 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a"} err="failed to get container status \"b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a\": rpc error: code = NotFound desc = could not find container \"b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a\": container with ID starting with b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a not found: ID does not exist" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.600461 4787 scope.go:117] "RemoveContainer" containerID="394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.600687 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694"} err="failed to get container status \"394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694\": rpc error: code = NotFound desc = could not find container \"394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694\": container with ID starting with 394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694 not found: ID does not exist" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.600714 4787 scope.go:117] "RemoveContainer" containerID="a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.600908 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c"} err="failed to get container status \"a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\": rpc error: code = NotFound desc = could not find container \"a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\": container with ID starting with a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c not found: ID does not exist" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.600924 4787 scope.go:117] "RemoveContainer" containerID="81f37f28240c3a86515abe10685e1a875a8cafa2df45d92b131677f3b1eacdac" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.601148 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81f37f28240c3a86515abe10685e1a875a8cafa2df45d92b131677f3b1eacdac"} err="failed to get container status \"81f37f28240c3a86515abe10685e1a875a8cafa2df45d92b131677f3b1eacdac\": rpc error: code = NotFound desc = could not find container \"81f37f28240c3a86515abe10685e1a875a8cafa2df45d92b131677f3b1eacdac\": container with ID starting with 81f37f28240c3a86515abe10685e1a875a8cafa2df45d92b131677f3b1eacdac not found: ID does not exist" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.601168 4787 scope.go:117] "RemoveContainer" containerID="c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.601371 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122"} err="failed to get container status \"c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122\": rpc error: code = NotFound desc = could not find container \"c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122\": container with ID starting with c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122 not found: ID does not exist" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.601396 4787 scope.go:117] "RemoveContainer" containerID="85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.601574 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68"} err="failed to get container status \"85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68\": rpc error: code = NotFound desc = could not find container \"85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68\": container with ID starting with 85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68 not found: ID does not exist" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.601596 4787 scope.go:117] "RemoveContainer" containerID="e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.601782 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4"} err="failed to get container status \"e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4\": rpc error: code = NotFound desc = could not find container \"e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4\": container with ID starting with e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4 not found: ID does not exist" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.601801 4787 scope.go:117] "RemoveContainer" containerID="834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.601982 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e"} err="failed to get container status \"834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e\": rpc error: code = NotFound desc = could not find container \"834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e\": container with ID starting with 834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e not found: ID does not exist" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.602003 4787 scope.go:117] "RemoveContainer" containerID="39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.602223 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274"} err="failed to get container status \"39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274\": rpc error: code = NotFound desc = could not find container \"39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274\": container with ID starting with 39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274 not found: ID does not exist" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.602244 4787 scope.go:117] "RemoveContainer" containerID="7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.602443 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f"} err="failed to get container status \"7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f\": rpc error: code = NotFound desc = could not find container \"7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f\": container with ID starting with 7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f not found: ID does not exist" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.602464 4787 scope.go:117] "RemoveContainer" containerID="b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.602647 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a"} err="failed to get container status \"b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a\": rpc error: code = NotFound desc = could not find container \"b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a\": container with ID starting with b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a not found: ID does not exist" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.602669 4787 scope.go:117] "RemoveContainer" containerID="394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.602853 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694"} err="failed to get container status \"394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694\": rpc error: code = NotFound desc = could not find container \"394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694\": container with ID starting with 394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694 not found: ID does not exist" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.602873 4787 scope.go:117] "RemoveContainer" containerID="a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.603066 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c"} err="failed to get container status \"a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\": rpc error: code = NotFound desc = could not find container \"a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\": container with ID starting with a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c not found: ID does not exist" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.603114 4787 scope.go:117] "RemoveContainer" containerID="81f37f28240c3a86515abe10685e1a875a8cafa2df45d92b131677f3b1eacdac" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.603285 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81f37f28240c3a86515abe10685e1a875a8cafa2df45d92b131677f3b1eacdac"} err="failed to get container status \"81f37f28240c3a86515abe10685e1a875a8cafa2df45d92b131677f3b1eacdac\": rpc error: code = NotFound desc = could not find container \"81f37f28240c3a86515abe10685e1a875a8cafa2df45d92b131677f3b1eacdac\": container with ID starting with 81f37f28240c3a86515abe10685e1a875a8cafa2df45d92b131677f3b1eacdac not found: ID does not exist" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.603305 4787 scope.go:117] "RemoveContainer" containerID="c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.603549 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122"} err="failed to get container status \"c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122\": rpc error: code = NotFound desc = could not find container \"c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122\": container with ID starting with c949148c152e77bac419bf78d51bf1ec3101f726d814fdd042aa06806abfd122 not found: ID does not exist" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.603565 4787 scope.go:117] "RemoveContainer" containerID="85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.604250 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68"} err="failed to get container status \"85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68\": rpc error: code = NotFound desc = could not find container \"85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68\": container with ID starting with 85aa4ad87e5f1cdbd7498e22c36c871b5978d581bb1913ac0188ded079059e68 not found: ID does not exist" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.604274 4787 scope.go:117] "RemoveContainer" containerID="e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.604805 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4"} err="failed to get container status \"e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4\": rpc error: code = NotFound desc = could not find container \"e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4\": container with ID starting with e7c54c9f19d9d1ee036120b4f728410a852d20ffbf4eb482a73b35dc1d7d20c4 not found: ID does not exist" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.604829 4787 scope.go:117] "RemoveContainer" containerID="834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.605120 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e"} err="failed to get container status \"834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e\": rpc error: code = NotFound desc = could not find container \"834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e\": container with ID starting with 834c3d3e9db31f11fcc111e184dee66f4ea6f9bf4901a693289629d001960d5e not found: ID does not exist" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.605142 4787 scope.go:117] "RemoveContainer" containerID="39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.605474 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274"} err="failed to get container status \"39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274\": rpc error: code = NotFound desc = could not find container \"39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274\": container with ID starting with 39444b78461dc0156c64a6ca55fcff3cdec6917283fb549d31ddaeb6ce03a274 not found: ID does not exist" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.605508 4787 scope.go:117] "RemoveContainer" containerID="7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.605948 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f"} err="failed to get container status \"7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f\": rpc error: code = NotFound desc = could not find container \"7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f\": container with ID starting with 7fd072ed908823e9c63f3f6db742d8c51934e36580063d9e38417d3b0f31fa5f not found: ID does not exist" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.605981 4787 scope.go:117] "RemoveContainer" containerID="b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.606234 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a"} err="failed to get container status \"b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a\": rpc error: code = NotFound desc = could not find container \"b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a\": container with ID starting with b57ebd14800c1b32290dc9e9ab33fdd16f08a35455d3625a68ca9eec68ed8a3a not found: ID does not exist" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.606265 4787 scope.go:117] "RemoveContainer" containerID="394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.606496 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694"} err="failed to get container status \"394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694\": rpc error: code = NotFound desc = could not find container \"394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694\": container with ID starting with 394fc6aa6d5d0888a8e88a2431b569a56bbee58a5abeae6c778092d9410a7694 not found: ID does not exist" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.606523 4787 scope.go:117] "RemoveContainer" containerID="a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c" Oct 01 09:45:54 crc kubenswrapper[4787]: I1001 09:45:54.606795 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c"} err="failed to get container status \"a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\": rpc error: code = NotFound desc = could not find container \"a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c\": container with ID starting with a69b102332a71693512dd08914dd63eba69c18ec6074a949a1d75c69e7d3b65c not found: ID does not exist" Oct 01 09:45:55 crc kubenswrapper[4787]: I1001 09:45:55.352814 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z8tjz_53f5ffd2-7649-4c9e-929f-a68187409644/kube-multus/2.log" Oct 01 09:45:55 crc kubenswrapper[4787]: I1001 09:45:55.359230 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" event={"ID":"29fc1929-bc7b-4340-afd7-f801f43909a1","Type":"ContainerStarted","Data":"073f7af3a0d93bbc4e84a0722015a293a369f9b198bc5abdce5fda09d5060838"} Oct 01 09:45:55 crc kubenswrapper[4787]: I1001 09:45:55.359313 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" event={"ID":"29fc1929-bc7b-4340-afd7-f801f43909a1","Type":"ContainerStarted","Data":"624a8723d05c276dbd268ce9478d809bf9fa94d8f25c46814a8fed71417e2c7c"} Oct 01 09:45:55 crc kubenswrapper[4787]: I1001 09:45:55.359326 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" event={"ID":"29fc1929-bc7b-4340-afd7-f801f43909a1","Type":"ContainerStarted","Data":"bd8f17703fc48dc19e301bafe7348fdec00b4a77d6bceed69d34d0d108f3f74d"} Oct 01 09:45:55 crc kubenswrapper[4787]: I1001 09:45:55.359342 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" event={"ID":"29fc1929-bc7b-4340-afd7-f801f43909a1","Type":"ContainerStarted","Data":"42be5ee50ed53587b07e2a9288c0e5e00ae1d288d0d6da735f17e0108180492a"} Oct 01 09:45:55 crc kubenswrapper[4787]: I1001 09:45:55.359354 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" event={"ID":"29fc1929-bc7b-4340-afd7-f801f43909a1","Type":"ContainerStarted","Data":"50c313c019f08164a357b7e725e7ffb4899f36e81195c2fc1e9b75fa93834d7c"} Oct 01 09:45:55 crc kubenswrapper[4787]: I1001 09:45:55.359365 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" event={"ID":"29fc1929-bc7b-4340-afd7-f801f43909a1","Type":"ContainerStarted","Data":"c3f2859d31fdc54fb94a9ee3bb26c2a6c55d209d68c95b443a14e074718bf8d7"} Oct 01 09:45:57 crc kubenswrapper[4787]: I1001 09:45:57.379140 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" event={"ID":"29fc1929-bc7b-4340-afd7-f801f43909a1","Type":"ContainerStarted","Data":"4a790d9ea080337a9747e0b6da6b4bc650765a752ccefedaac1021777575211c"} Oct 01 09:46:00 crc kubenswrapper[4787]: I1001 09:46:00.403428 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" event={"ID":"29fc1929-bc7b-4340-afd7-f801f43909a1","Type":"ContainerStarted","Data":"9fb4708c60e4b7da9b86bbb0e8677e09ab0e29c66ca9a8dfaa326b192af879ad"} Oct 01 09:46:00 crc kubenswrapper[4787]: I1001 09:46:00.404022 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:46:00 crc kubenswrapper[4787]: I1001 09:46:00.404039 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:46:00 crc kubenswrapper[4787]: I1001 09:46:00.404050 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:46:00 crc kubenswrapper[4787]: I1001 09:46:00.433901 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:46:00 crc kubenswrapper[4787]: I1001 09:46:00.446402 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:46:00 crc kubenswrapper[4787]: I1001 09:46:00.447284 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" podStartSLOduration=7.447266395 podStartE2EDuration="7.447266395s" podCreationTimestamp="2025-10-01 09:45:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:46:00.446036563 +0000 UTC m=+592.561180720" watchObservedRunningTime="2025-10-01 09:46:00.447266395 +0000 UTC m=+592.562410572" Oct 01 09:46:05 crc kubenswrapper[4787]: I1001 09:46:05.523996 4787 scope.go:117] "RemoveContainer" containerID="8ccabd0db928904435aacf77b37e997bf489a08e4cc113b3438d986e0c7f0ed2" Oct 01 09:46:05 crc kubenswrapper[4787]: E1001 09:46:05.524797 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-z8tjz_openshift-multus(53f5ffd2-7649-4c9e-929f-a68187409644)\"" pod="openshift-multus/multus-z8tjz" podUID="53f5ffd2-7649-4c9e-929f-a68187409644" Oct 01 09:46:20 crc kubenswrapper[4787]: I1001 09:46:20.524179 4787 scope.go:117] "RemoveContainer" containerID="8ccabd0db928904435aacf77b37e997bf489a08e4cc113b3438d986e0c7f0ed2" Oct 01 09:46:21 crc kubenswrapper[4787]: I1001 09:46:21.520774 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z8tjz_53f5ffd2-7649-4c9e-929f-a68187409644/kube-multus/2.log" Oct 01 09:46:21 crc kubenswrapper[4787]: I1001 09:46:21.521780 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-z8tjz" event={"ID":"53f5ffd2-7649-4c9e-929f-a68187409644","Type":"ContainerStarted","Data":"5be9c9e1e39bca1b6327590f13652475cb8b9a049d0d20cc4109fafc0ddd318f"} Oct 01 09:46:24 crc kubenswrapper[4787]: I1001 09:46:24.093865 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-vv8p9" Oct 01 09:46:30 crc kubenswrapper[4787]: I1001 09:46:30.830937 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc"] Oct 01 09:46:30 crc kubenswrapper[4787]: I1001 09:46:30.833485 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc" Oct 01 09:46:30 crc kubenswrapper[4787]: I1001 09:46:30.836212 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 01 09:46:30 crc kubenswrapper[4787]: I1001 09:46:30.848436 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc"] Oct 01 09:46:30 crc kubenswrapper[4787]: I1001 09:46:30.930214 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e60e8301-91e9-45d5-a3d1-cd29ce1e9514-util\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc\" (UID: \"e60e8301-91e9-45d5-a3d1-cd29ce1e9514\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc" Oct 01 09:46:30 crc kubenswrapper[4787]: I1001 09:46:30.930290 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrcs2\" (UniqueName: \"kubernetes.io/projected/e60e8301-91e9-45d5-a3d1-cd29ce1e9514-kube-api-access-wrcs2\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc\" (UID: \"e60e8301-91e9-45d5-a3d1-cd29ce1e9514\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc" Oct 01 09:46:30 crc kubenswrapper[4787]: I1001 09:46:30.930328 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e60e8301-91e9-45d5-a3d1-cd29ce1e9514-bundle\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc\" (UID: \"e60e8301-91e9-45d5-a3d1-cd29ce1e9514\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc" Oct 01 09:46:31 crc kubenswrapper[4787]: I1001 09:46:31.032283 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e60e8301-91e9-45d5-a3d1-cd29ce1e9514-util\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc\" (UID: \"e60e8301-91e9-45d5-a3d1-cd29ce1e9514\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc" Oct 01 09:46:31 crc kubenswrapper[4787]: I1001 09:46:31.032412 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrcs2\" (UniqueName: \"kubernetes.io/projected/e60e8301-91e9-45d5-a3d1-cd29ce1e9514-kube-api-access-wrcs2\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc\" (UID: \"e60e8301-91e9-45d5-a3d1-cd29ce1e9514\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc" Oct 01 09:46:31 crc kubenswrapper[4787]: I1001 09:46:31.032466 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e60e8301-91e9-45d5-a3d1-cd29ce1e9514-bundle\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc\" (UID: \"e60e8301-91e9-45d5-a3d1-cd29ce1e9514\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc" Oct 01 09:46:31 crc kubenswrapper[4787]: I1001 09:46:31.033726 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e60e8301-91e9-45d5-a3d1-cd29ce1e9514-bundle\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc\" (UID: \"e60e8301-91e9-45d5-a3d1-cd29ce1e9514\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc" Oct 01 09:46:31 crc kubenswrapper[4787]: I1001 09:46:31.034543 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e60e8301-91e9-45d5-a3d1-cd29ce1e9514-util\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc\" (UID: \"e60e8301-91e9-45d5-a3d1-cd29ce1e9514\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc" Oct 01 09:46:31 crc kubenswrapper[4787]: I1001 09:46:31.061536 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrcs2\" (UniqueName: \"kubernetes.io/projected/e60e8301-91e9-45d5-a3d1-cd29ce1e9514-kube-api-access-wrcs2\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc\" (UID: \"e60e8301-91e9-45d5-a3d1-cd29ce1e9514\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc" Oct 01 09:46:31 crc kubenswrapper[4787]: I1001 09:46:31.155011 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc" Oct 01 09:46:31 crc kubenswrapper[4787]: I1001 09:46:31.406462 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc"] Oct 01 09:46:31 crc kubenswrapper[4787]: I1001 09:46:31.601112 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc" event={"ID":"e60e8301-91e9-45d5-a3d1-cd29ce1e9514","Type":"ContainerStarted","Data":"f1b1451589f6b960159a1408c36816ed979bf2bf7c1263a0ed889e27d8becfb6"} Oct 01 09:46:31 crc kubenswrapper[4787]: I1001 09:46:31.601165 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc" event={"ID":"e60e8301-91e9-45d5-a3d1-cd29ce1e9514","Type":"ContainerStarted","Data":"53a78ff8c56109115cb856d8084f9840282e3e5ac7d98e350dfc1286d1b9c515"} Oct 01 09:46:32 crc kubenswrapper[4787]: I1001 09:46:32.609660 4787 generic.go:334] "Generic (PLEG): container finished" podID="e60e8301-91e9-45d5-a3d1-cd29ce1e9514" containerID="f1b1451589f6b960159a1408c36816ed979bf2bf7c1263a0ed889e27d8becfb6" exitCode=0 Oct 01 09:46:32 crc kubenswrapper[4787]: I1001 09:46:32.609735 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc" event={"ID":"e60e8301-91e9-45d5-a3d1-cd29ce1e9514","Type":"ContainerDied","Data":"f1b1451589f6b960159a1408c36816ed979bf2bf7c1263a0ed889e27d8becfb6"} Oct 01 09:46:34 crc kubenswrapper[4787]: I1001 09:46:34.636064 4787 generic.go:334] "Generic (PLEG): container finished" podID="e60e8301-91e9-45d5-a3d1-cd29ce1e9514" containerID="660b7bed1ed1131b7f7c3afe94b365b924ae01ab4d800f676243f4e23e8ddc1a" exitCode=0 Oct 01 09:46:34 crc kubenswrapper[4787]: I1001 09:46:34.636140 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc" event={"ID":"e60e8301-91e9-45d5-a3d1-cd29ce1e9514","Type":"ContainerDied","Data":"660b7bed1ed1131b7f7c3afe94b365b924ae01ab4d800f676243f4e23e8ddc1a"} Oct 01 09:46:35 crc kubenswrapper[4787]: I1001 09:46:35.650540 4787 generic.go:334] "Generic (PLEG): container finished" podID="e60e8301-91e9-45d5-a3d1-cd29ce1e9514" containerID="7aba287e446a215731d7845899ccd36df8492027fc468d6121fc699a7b874170" exitCode=0 Oct 01 09:46:35 crc kubenswrapper[4787]: I1001 09:46:35.650624 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc" event={"ID":"e60e8301-91e9-45d5-a3d1-cd29ce1e9514","Type":"ContainerDied","Data":"7aba287e446a215731d7845899ccd36df8492027fc468d6121fc699a7b874170"} Oct 01 09:46:37 crc kubenswrapper[4787]: I1001 09:46:37.008892 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc" Oct 01 09:46:37 crc kubenswrapper[4787]: I1001 09:46:37.125508 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e60e8301-91e9-45d5-a3d1-cd29ce1e9514-bundle\") pod \"e60e8301-91e9-45d5-a3d1-cd29ce1e9514\" (UID: \"e60e8301-91e9-45d5-a3d1-cd29ce1e9514\") " Oct 01 09:46:37 crc kubenswrapper[4787]: I1001 09:46:37.125595 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e60e8301-91e9-45d5-a3d1-cd29ce1e9514-util\") pod \"e60e8301-91e9-45d5-a3d1-cd29ce1e9514\" (UID: \"e60e8301-91e9-45d5-a3d1-cd29ce1e9514\") " Oct 01 09:46:37 crc kubenswrapper[4787]: I1001 09:46:37.125713 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrcs2\" (UniqueName: \"kubernetes.io/projected/e60e8301-91e9-45d5-a3d1-cd29ce1e9514-kube-api-access-wrcs2\") pod \"e60e8301-91e9-45d5-a3d1-cd29ce1e9514\" (UID: \"e60e8301-91e9-45d5-a3d1-cd29ce1e9514\") " Oct 01 09:46:37 crc kubenswrapper[4787]: I1001 09:46:37.126800 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e60e8301-91e9-45d5-a3d1-cd29ce1e9514-bundle" (OuterVolumeSpecName: "bundle") pod "e60e8301-91e9-45d5-a3d1-cd29ce1e9514" (UID: "e60e8301-91e9-45d5-a3d1-cd29ce1e9514"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:46:37 crc kubenswrapper[4787]: I1001 09:46:37.138203 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e60e8301-91e9-45d5-a3d1-cd29ce1e9514-kube-api-access-wrcs2" (OuterVolumeSpecName: "kube-api-access-wrcs2") pod "e60e8301-91e9-45d5-a3d1-cd29ce1e9514" (UID: "e60e8301-91e9-45d5-a3d1-cd29ce1e9514"). InnerVolumeSpecName "kube-api-access-wrcs2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:46:37 crc kubenswrapper[4787]: I1001 09:46:37.228208 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrcs2\" (UniqueName: \"kubernetes.io/projected/e60e8301-91e9-45d5-a3d1-cd29ce1e9514-kube-api-access-wrcs2\") on node \"crc\" DevicePath \"\"" Oct 01 09:46:37 crc kubenswrapper[4787]: I1001 09:46:37.228282 4787 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e60e8301-91e9-45d5-a3d1-cd29ce1e9514-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:46:37 crc kubenswrapper[4787]: I1001 09:46:37.381734 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e60e8301-91e9-45d5-a3d1-cd29ce1e9514-util" (OuterVolumeSpecName: "util") pod "e60e8301-91e9-45d5-a3d1-cd29ce1e9514" (UID: "e60e8301-91e9-45d5-a3d1-cd29ce1e9514"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:46:37 crc kubenswrapper[4787]: I1001 09:46:37.431410 4787 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e60e8301-91e9-45d5-a3d1-cd29ce1e9514-util\") on node \"crc\" DevicePath \"\"" Oct 01 09:46:37 crc kubenswrapper[4787]: I1001 09:46:37.669313 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc" event={"ID":"e60e8301-91e9-45d5-a3d1-cd29ce1e9514","Type":"ContainerDied","Data":"53a78ff8c56109115cb856d8084f9840282e3e5ac7d98e350dfc1286d1b9c515"} Oct 01 09:46:37 crc kubenswrapper[4787]: I1001 09:46:37.669360 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="53a78ff8c56109115cb856d8084f9840282e3e5ac7d98e350dfc1286d1b9c515" Oct 01 09:46:37 crc kubenswrapper[4787]: I1001 09:46:37.669467 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc" Oct 01 09:46:39 crc kubenswrapper[4787]: I1001 09:46:39.899403 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5d6f6cfd66-8l5nr"] Oct 01 09:46:39 crc kubenswrapper[4787]: E1001 09:46:39.900316 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e60e8301-91e9-45d5-a3d1-cd29ce1e9514" containerName="extract" Oct 01 09:46:39 crc kubenswrapper[4787]: I1001 09:46:39.900340 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e60e8301-91e9-45d5-a3d1-cd29ce1e9514" containerName="extract" Oct 01 09:46:39 crc kubenswrapper[4787]: E1001 09:46:39.900369 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e60e8301-91e9-45d5-a3d1-cd29ce1e9514" containerName="util" Oct 01 09:46:39 crc kubenswrapper[4787]: I1001 09:46:39.900378 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e60e8301-91e9-45d5-a3d1-cd29ce1e9514" containerName="util" Oct 01 09:46:39 crc kubenswrapper[4787]: E1001 09:46:39.900388 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e60e8301-91e9-45d5-a3d1-cd29ce1e9514" containerName="pull" Oct 01 09:46:39 crc kubenswrapper[4787]: I1001 09:46:39.900399 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e60e8301-91e9-45d5-a3d1-cd29ce1e9514" containerName="pull" Oct 01 09:46:39 crc kubenswrapper[4787]: I1001 09:46:39.900553 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="e60e8301-91e9-45d5-a3d1-cd29ce1e9514" containerName="extract" Oct 01 09:46:39 crc kubenswrapper[4787]: I1001 09:46:39.901217 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-8l5nr" Oct 01 09:46:39 crc kubenswrapper[4787]: I1001 09:46:39.903479 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 01 09:46:39 crc kubenswrapper[4787]: I1001 09:46:39.903651 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-bnl2w" Oct 01 09:46:39 crc kubenswrapper[4787]: I1001 09:46:39.905201 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 01 09:46:39 crc kubenswrapper[4787]: I1001 09:46:39.913874 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5d6f6cfd66-8l5nr"] Oct 01 09:46:39 crc kubenswrapper[4787]: I1001 09:46:39.967795 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqvzt\" (UniqueName: \"kubernetes.io/projected/27ab4c98-35a2-45be-864e-4550be0fa875-kube-api-access-mqvzt\") pod \"nmstate-operator-5d6f6cfd66-8l5nr\" (UID: \"27ab4c98-35a2-45be-864e-4550be0fa875\") " pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-8l5nr" Oct 01 09:46:40 crc kubenswrapper[4787]: I1001 09:46:40.068672 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqvzt\" (UniqueName: \"kubernetes.io/projected/27ab4c98-35a2-45be-864e-4550be0fa875-kube-api-access-mqvzt\") pod \"nmstate-operator-5d6f6cfd66-8l5nr\" (UID: \"27ab4c98-35a2-45be-864e-4550be0fa875\") " pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-8l5nr" Oct 01 09:46:40 crc kubenswrapper[4787]: I1001 09:46:40.088224 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqvzt\" (UniqueName: \"kubernetes.io/projected/27ab4c98-35a2-45be-864e-4550be0fa875-kube-api-access-mqvzt\") pod \"nmstate-operator-5d6f6cfd66-8l5nr\" (UID: \"27ab4c98-35a2-45be-864e-4550be0fa875\") " pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-8l5nr" Oct 01 09:46:40 crc kubenswrapper[4787]: I1001 09:46:40.221817 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-8l5nr" Oct 01 09:46:40 crc kubenswrapper[4787]: I1001 09:46:40.469418 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5d6f6cfd66-8l5nr"] Oct 01 09:46:40 crc kubenswrapper[4787]: I1001 09:46:40.688807 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-8l5nr" event={"ID":"27ab4c98-35a2-45be-864e-4550be0fa875","Type":"ContainerStarted","Data":"e8b73e6bb354c86d486223074df19f49a1575db89aaa0fcc77b62419b0289044"} Oct 01 09:46:43 crc kubenswrapper[4787]: I1001 09:46:43.709616 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-8l5nr" event={"ID":"27ab4c98-35a2-45be-864e-4550be0fa875","Type":"ContainerStarted","Data":"bc269cb14cc749330cdd17f2bf20f6cfec310cdffd19b264614919e2d100e4d8"} Oct 01 09:46:43 crc kubenswrapper[4787]: I1001 09:46:43.729195 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-8l5nr" podStartSLOduration=2.600316017 podStartE2EDuration="4.729179552s" podCreationTimestamp="2025-10-01 09:46:39 +0000 UTC" firstStartedPulling="2025-10-01 09:46:40.479046944 +0000 UTC m=+632.594191101" lastFinishedPulling="2025-10-01 09:46:42.607910479 +0000 UTC m=+634.723054636" observedRunningTime="2025-10-01 09:46:43.727067767 +0000 UTC m=+635.842211924" watchObservedRunningTime="2025-10-01 09:46:43.729179552 +0000 UTC m=+635.844323709" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.059893 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-58fcddf996-2cnst"] Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.062057 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58fcddf996-2cnst" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.064319 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6d689559c5-22sp6"] Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.065413 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6d689559c5-22sp6" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.066284 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-b8z2t" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.067828 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.087901 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58fcddf996-2cnst"] Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.093268 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-2rzr2"] Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.094237 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-2rzr2" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.103181 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6d689559c5-22sp6"] Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.136700 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkpsd\" (UniqueName: \"kubernetes.io/projected/ee9e051d-7391-4a43-9a02-5084da577b36-kube-api-access-mkpsd\") pod \"nmstate-handler-2rzr2\" (UID: \"ee9e051d-7391-4a43-9a02-5084da577b36\") " pod="openshift-nmstate/nmstate-handler-2rzr2" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.136768 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/ee9e051d-7391-4a43-9a02-5084da577b36-dbus-socket\") pod \"nmstate-handler-2rzr2\" (UID: \"ee9e051d-7391-4a43-9a02-5084da577b36\") " pod="openshift-nmstate/nmstate-handler-2rzr2" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.136796 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/ee9e051d-7391-4a43-9a02-5084da577b36-nmstate-lock\") pod \"nmstate-handler-2rzr2\" (UID: \"ee9e051d-7391-4a43-9a02-5084da577b36\") " pod="openshift-nmstate/nmstate-handler-2rzr2" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.136838 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvd9z\" (UniqueName: \"kubernetes.io/projected/97efd9a6-e221-4aac-97b1-b49de45edd88-kube-api-access-xvd9z\") pod \"nmstate-webhook-6d689559c5-22sp6\" (UID: \"97efd9a6-e221-4aac-97b1-b49de45edd88\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-22sp6" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.136861 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/97efd9a6-e221-4aac-97b1-b49de45edd88-tls-key-pair\") pod \"nmstate-webhook-6d689559c5-22sp6\" (UID: \"97efd9a6-e221-4aac-97b1-b49de45edd88\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-22sp6" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.136962 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ggs7\" (UniqueName: \"kubernetes.io/projected/10cd7a64-357c-4c0d-b11c-426e4060b61c-kube-api-access-7ggs7\") pod \"nmstate-metrics-58fcddf996-2cnst\" (UID: \"10cd7a64-357c-4c0d-b11c-426e4060b61c\") " pod="openshift-nmstate/nmstate-metrics-58fcddf996-2cnst" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.137031 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/ee9e051d-7391-4a43-9a02-5084da577b36-ovs-socket\") pod \"nmstate-handler-2rzr2\" (UID: \"ee9e051d-7391-4a43-9a02-5084da577b36\") " pod="openshift-nmstate/nmstate-handler-2rzr2" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.210364 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-864bb6dfb5-cjhqm"] Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.211217 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-cjhqm" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.213300 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.215063 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-zsjw8" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.215445 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.224421 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-864bb6dfb5-cjhqm"] Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.237913 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/e4a805b1-d488-4007-9dd3-38d3fbab1f10-nginx-conf\") pod \"nmstate-console-plugin-864bb6dfb5-cjhqm\" (UID: \"e4a805b1-d488-4007-9dd3-38d3fbab1f10\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-cjhqm" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.237958 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/e4a805b1-d488-4007-9dd3-38d3fbab1f10-plugin-serving-cert\") pod \"nmstate-console-plugin-864bb6dfb5-cjhqm\" (UID: \"e4a805b1-d488-4007-9dd3-38d3fbab1f10\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-cjhqm" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.237993 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/ee9e051d-7391-4a43-9a02-5084da577b36-ovs-socket\") pod \"nmstate-handler-2rzr2\" (UID: \"ee9e051d-7391-4a43-9a02-5084da577b36\") " pod="openshift-nmstate/nmstate-handler-2rzr2" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.238041 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/ee9e051d-7391-4a43-9a02-5084da577b36-ovs-socket\") pod \"nmstate-handler-2rzr2\" (UID: \"ee9e051d-7391-4a43-9a02-5084da577b36\") " pod="openshift-nmstate/nmstate-handler-2rzr2" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.238297 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkpsd\" (UniqueName: \"kubernetes.io/projected/ee9e051d-7391-4a43-9a02-5084da577b36-kube-api-access-mkpsd\") pod \"nmstate-handler-2rzr2\" (UID: \"ee9e051d-7391-4a43-9a02-5084da577b36\") " pod="openshift-nmstate/nmstate-handler-2rzr2" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.238355 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/ee9e051d-7391-4a43-9a02-5084da577b36-dbus-socket\") pod \"nmstate-handler-2rzr2\" (UID: \"ee9e051d-7391-4a43-9a02-5084da577b36\") " pod="openshift-nmstate/nmstate-handler-2rzr2" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.238384 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/ee9e051d-7391-4a43-9a02-5084da577b36-nmstate-lock\") pod \"nmstate-handler-2rzr2\" (UID: \"ee9e051d-7391-4a43-9a02-5084da577b36\") " pod="openshift-nmstate/nmstate-handler-2rzr2" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.238443 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/97efd9a6-e221-4aac-97b1-b49de45edd88-tls-key-pair\") pod \"nmstate-webhook-6d689559c5-22sp6\" (UID: \"97efd9a6-e221-4aac-97b1-b49de45edd88\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-22sp6" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.238464 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvd9z\" (UniqueName: \"kubernetes.io/projected/97efd9a6-e221-4aac-97b1-b49de45edd88-kube-api-access-xvd9z\") pod \"nmstate-webhook-6d689559c5-22sp6\" (UID: \"97efd9a6-e221-4aac-97b1-b49de45edd88\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-22sp6" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.238523 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2r8w\" (UniqueName: \"kubernetes.io/projected/e4a805b1-d488-4007-9dd3-38d3fbab1f10-kube-api-access-q2r8w\") pod \"nmstate-console-plugin-864bb6dfb5-cjhqm\" (UID: \"e4a805b1-d488-4007-9dd3-38d3fbab1f10\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-cjhqm" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.238589 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/ee9e051d-7391-4a43-9a02-5084da577b36-nmstate-lock\") pod \"nmstate-handler-2rzr2\" (UID: \"ee9e051d-7391-4a43-9a02-5084da577b36\") " pod="openshift-nmstate/nmstate-handler-2rzr2" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.238656 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ggs7\" (UniqueName: \"kubernetes.io/projected/10cd7a64-357c-4c0d-b11c-426e4060b61c-kube-api-access-7ggs7\") pod \"nmstate-metrics-58fcddf996-2cnst\" (UID: \"10cd7a64-357c-4c0d-b11c-426e4060b61c\") " pod="openshift-nmstate/nmstate-metrics-58fcddf996-2cnst" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.238674 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/ee9e051d-7391-4a43-9a02-5084da577b36-dbus-socket\") pod \"nmstate-handler-2rzr2\" (UID: \"ee9e051d-7391-4a43-9a02-5084da577b36\") " pod="openshift-nmstate/nmstate-handler-2rzr2" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.260517 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/97efd9a6-e221-4aac-97b1-b49de45edd88-tls-key-pair\") pod \"nmstate-webhook-6d689559c5-22sp6\" (UID: \"97efd9a6-e221-4aac-97b1-b49de45edd88\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-22sp6" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.264225 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvd9z\" (UniqueName: \"kubernetes.io/projected/97efd9a6-e221-4aac-97b1-b49de45edd88-kube-api-access-xvd9z\") pod \"nmstate-webhook-6d689559c5-22sp6\" (UID: \"97efd9a6-e221-4aac-97b1-b49de45edd88\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-22sp6" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.264328 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkpsd\" (UniqueName: \"kubernetes.io/projected/ee9e051d-7391-4a43-9a02-5084da577b36-kube-api-access-mkpsd\") pod \"nmstate-handler-2rzr2\" (UID: \"ee9e051d-7391-4a43-9a02-5084da577b36\") " pod="openshift-nmstate/nmstate-handler-2rzr2" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.268470 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ggs7\" (UniqueName: \"kubernetes.io/projected/10cd7a64-357c-4c0d-b11c-426e4060b61c-kube-api-access-7ggs7\") pod \"nmstate-metrics-58fcddf996-2cnst\" (UID: \"10cd7a64-357c-4c0d-b11c-426e4060b61c\") " pod="openshift-nmstate/nmstate-metrics-58fcddf996-2cnst" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.339897 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2r8w\" (UniqueName: \"kubernetes.io/projected/e4a805b1-d488-4007-9dd3-38d3fbab1f10-kube-api-access-q2r8w\") pod \"nmstate-console-plugin-864bb6dfb5-cjhqm\" (UID: \"e4a805b1-d488-4007-9dd3-38d3fbab1f10\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-cjhqm" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.339970 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/e4a805b1-d488-4007-9dd3-38d3fbab1f10-nginx-conf\") pod \"nmstate-console-plugin-864bb6dfb5-cjhqm\" (UID: \"e4a805b1-d488-4007-9dd3-38d3fbab1f10\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-cjhqm" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.339992 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/e4a805b1-d488-4007-9dd3-38d3fbab1f10-plugin-serving-cert\") pod \"nmstate-console-plugin-864bb6dfb5-cjhqm\" (UID: \"e4a805b1-d488-4007-9dd3-38d3fbab1f10\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-cjhqm" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.341519 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/e4a805b1-d488-4007-9dd3-38d3fbab1f10-nginx-conf\") pod \"nmstate-console-plugin-864bb6dfb5-cjhqm\" (UID: \"e4a805b1-d488-4007-9dd3-38d3fbab1f10\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-cjhqm" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.354646 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/e4a805b1-d488-4007-9dd3-38d3fbab1f10-plugin-serving-cert\") pod \"nmstate-console-plugin-864bb6dfb5-cjhqm\" (UID: \"e4a805b1-d488-4007-9dd3-38d3fbab1f10\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-cjhqm" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.361565 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2r8w\" (UniqueName: \"kubernetes.io/projected/e4a805b1-d488-4007-9dd3-38d3fbab1f10-kube-api-access-q2r8w\") pod \"nmstate-console-plugin-864bb6dfb5-cjhqm\" (UID: \"e4a805b1-d488-4007-9dd3-38d3fbab1f10\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-cjhqm" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.385234 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58fcddf996-2cnst" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.394502 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6d689559c5-22sp6" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.410252 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-2rzr2" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.426859 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-5c65595bb4-dvx6f"] Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.427687 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5c65595bb4-dvx6f" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.444770 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5c65595bb4-dvx6f"] Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.535935 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-cjhqm" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.545622 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2e57c36d-b34a-43eb-8c96-15956f1bfacd-trusted-ca-bundle\") pod \"console-5c65595bb4-dvx6f\" (UID: \"2e57c36d-b34a-43eb-8c96-15956f1bfacd\") " pod="openshift-console/console-5c65595bb4-dvx6f" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.545675 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2e57c36d-b34a-43eb-8c96-15956f1bfacd-service-ca\") pod \"console-5c65595bb4-dvx6f\" (UID: \"2e57c36d-b34a-43eb-8c96-15956f1bfacd\") " pod="openshift-console/console-5c65595bb4-dvx6f" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.545722 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2e57c36d-b34a-43eb-8c96-15956f1bfacd-console-oauth-config\") pod \"console-5c65595bb4-dvx6f\" (UID: \"2e57c36d-b34a-43eb-8c96-15956f1bfacd\") " pod="openshift-console/console-5c65595bb4-dvx6f" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.545755 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2e57c36d-b34a-43eb-8c96-15956f1bfacd-oauth-serving-cert\") pod \"console-5c65595bb4-dvx6f\" (UID: \"2e57c36d-b34a-43eb-8c96-15956f1bfacd\") " pod="openshift-console/console-5c65595bb4-dvx6f" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.545782 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2e57c36d-b34a-43eb-8c96-15956f1bfacd-console-config\") pod \"console-5c65595bb4-dvx6f\" (UID: \"2e57c36d-b34a-43eb-8c96-15956f1bfacd\") " pod="openshift-console/console-5c65595bb4-dvx6f" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.545801 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7bpn\" (UniqueName: \"kubernetes.io/projected/2e57c36d-b34a-43eb-8c96-15956f1bfacd-kube-api-access-r7bpn\") pod \"console-5c65595bb4-dvx6f\" (UID: \"2e57c36d-b34a-43eb-8c96-15956f1bfacd\") " pod="openshift-console/console-5c65595bb4-dvx6f" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.545818 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2e57c36d-b34a-43eb-8c96-15956f1bfacd-console-serving-cert\") pod \"console-5c65595bb4-dvx6f\" (UID: \"2e57c36d-b34a-43eb-8c96-15956f1bfacd\") " pod="openshift-console/console-5c65595bb4-dvx6f" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.640124 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6d689559c5-22sp6"] Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.647290 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2e57c36d-b34a-43eb-8c96-15956f1bfacd-service-ca\") pod \"console-5c65595bb4-dvx6f\" (UID: \"2e57c36d-b34a-43eb-8c96-15956f1bfacd\") " pod="openshift-console/console-5c65595bb4-dvx6f" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.647352 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2e57c36d-b34a-43eb-8c96-15956f1bfacd-console-oauth-config\") pod \"console-5c65595bb4-dvx6f\" (UID: \"2e57c36d-b34a-43eb-8c96-15956f1bfacd\") " pod="openshift-console/console-5c65595bb4-dvx6f" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.647390 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2e57c36d-b34a-43eb-8c96-15956f1bfacd-oauth-serving-cert\") pod \"console-5c65595bb4-dvx6f\" (UID: \"2e57c36d-b34a-43eb-8c96-15956f1bfacd\") " pod="openshift-console/console-5c65595bb4-dvx6f" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.647408 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2e57c36d-b34a-43eb-8c96-15956f1bfacd-console-config\") pod \"console-5c65595bb4-dvx6f\" (UID: \"2e57c36d-b34a-43eb-8c96-15956f1bfacd\") " pod="openshift-console/console-5c65595bb4-dvx6f" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.647428 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7bpn\" (UniqueName: \"kubernetes.io/projected/2e57c36d-b34a-43eb-8c96-15956f1bfacd-kube-api-access-r7bpn\") pod \"console-5c65595bb4-dvx6f\" (UID: \"2e57c36d-b34a-43eb-8c96-15956f1bfacd\") " pod="openshift-console/console-5c65595bb4-dvx6f" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.647443 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2e57c36d-b34a-43eb-8c96-15956f1bfacd-console-serving-cert\") pod \"console-5c65595bb4-dvx6f\" (UID: \"2e57c36d-b34a-43eb-8c96-15956f1bfacd\") " pod="openshift-console/console-5c65595bb4-dvx6f" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.647504 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2e57c36d-b34a-43eb-8c96-15956f1bfacd-trusted-ca-bundle\") pod \"console-5c65595bb4-dvx6f\" (UID: \"2e57c36d-b34a-43eb-8c96-15956f1bfacd\") " pod="openshift-console/console-5c65595bb4-dvx6f" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.648829 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2e57c36d-b34a-43eb-8c96-15956f1bfacd-trusted-ca-bundle\") pod \"console-5c65595bb4-dvx6f\" (UID: \"2e57c36d-b34a-43eb-8c96-15956f1bfacd\") " pod="openshift-console/console-5c65595bb4-dvx6f" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.648842 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2e57c36d-b34a-43eb-8c96-15956f1bfacd-console-config\") pod \"console-5c65595bb4-dvx6f\" (UID: \"2e57c36d-b34a-43eb-8c96-15956f1bfacd\") " pod="openshift-console/console-5c65595bb4-dvx6f" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.649748 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2e57c36d-b34a-43eb-8c96-15956f1bfacd-service-ca\") pod \"console-5c65595bb4-dvx6f\" (UID: \"2e57c36d-b34a-43eb-8c96-15956f1bfacd\") " pod="openshift-console/console-5c65595bb4-dvx6f" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.650283 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2e57c36d-b34a-43eb-8c96-15956f1bfacd-oauth-serving-cert\") pod \"console-5c65595bb4-dvx6f\" (UID: \"2e57c36d-b34a-43eb-8c96-15956f1bfacd\") " pod="openshift-console/console-5c65595bb4-dvx6f" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.656647 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2e57c36d-b34a-43eb-8c96-15956f1bfacd-console-oauth-config\") pod \"console-5c65595bb4-dvx6f\" (UID: \"2e57c36d-b34a-43eb-8c96-15956f1bfacd\") " pod="openshift-console/console-5c65595bb4-dvx6f" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.658339 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2e57c36d-b34a-43eb-8c96-15956f1bfacd-console-serving-cert\") pod \"console-5c65595bb4-dvx6f\" (UID: \"2e57c36d-b34a-43eb-8c96-15956f1bfacd\") " pod="openshift-console/console-5c65595bb4-dvx6f" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.667720 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7bpn\" (UniqueName: \"kubernetes.io/projected/2e57c36d-b34a-43eb-8c96-15956f1bfacd-kube-api-access-r7bpn\") pod \"console-5c65595bb4-dvx6f\" (UID: \"2e57c36d-b34a-43eb-8c96-15956f1bfacd\") " pod="openshift-console/console-5c65595bb4-dvx6f" Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.751662 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-864bb6dfb5-cjhqm"] Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.752432 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5c65595bb4-dvx6f" Oct 01 09:46:51 crc kubenswrapper[4787]: W1001 09:46:51.757672 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode4a805b1_d488_4007_9dd3_38d3fbab1f10.slice/crio-7ebc4a5edb85c10a6abc8ef74825ec34142448d693b00b60bb92177c20fcf2f1 WatchSource:0}: Error finding container 7ebc4a5edb85c10a6abc8ef74825ec34142448d693b00b60bb92177c20fcf2f1: Status 404 returned error can't find the container with id 7ebc4a5edb85c10a6abc8ef74825ec34142448d693b00b60bb92177c20fcf2f1 Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.769025 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-cjhqm" event={"ID":"e4a805b1-d488-4007-9dd3-38d3fbab1f10","Type":"ContainerStarted","Data":"7ebc4a5edb85c10a6abc8ef74825ec34142448d693b00b60bb92177c20fcf2f1"} Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.770825 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-2rzr2" event={"ID":"ee9e051d-7391-4a43-9a02-5084da577b36","Type":"ContainerStarted","Data":"c4317c8cec688ba822ce86ae07babc906af51407b5d1cf780e9d6d176683fa42"} Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.772267 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6d689559c5-22sp6" event={"ID":"97efd9a6-e221-4aac-97b1-b49de45edd88","Type":"ContainerStarted","Data":"936e63c2c0e45ed924a8e4df9957b35928fe2cddc9ed097732836e5231cb4d7e"} Oct 01 09:46:51 crc kubenswrapper[4787]: I1001 09:46:51.884943 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58fcddf996-2cnst"] Oct 01 09:46:52 crc kubenswrapper[4787]: I1001 09:46:52.163332 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5c65595bb4-dvx6f"] Oct 01 09:46:52 crc kubenswrapper[4787]: W1001 09:46:52.171039 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2e57c36d_b34a_43eb_8c96_15956f1bfacd.slice/crio-51cf5877e51149692fb32be657f78c8d7c7d37378916d0c545861c3833eeb2f0 WatchSource:0}: Error finding container 51cf5877e51149692fb32be657f78c8d7c7d37378916d0c545861c3833eeb2f0: Status 404 returned error can't find the container with id 51cf5877e51149692fb32be657f78c8d7c7d37378916d0c545861c3833eeb2f0 Oct 01 09:46:52 crc kubenswrapper[4787]: I1001 09:46:52.781776 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58fcddf996-2cnst" event={"ID":"10cd7a64-357c-4c0d-b11c-426e4060b61c","Type":"ContainerStarted","Data":"3df09a66db7157279388d5a5735fdf2017af6a938fa0f3d024a62e5ff4fcbc4b"} Oct 01 09:46:52 crc kubenswrapper[4787]: I1001 09:46:52.785556 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5c65595bb4-dvx6f" event={"ID":"2e57c36d-b34a-43eb-8c96-15956f1bfacd","Type":"ContainerStarted","Data":"4b16616a7687fbc2a17850e70881d65904ce2c539965fc0c8e07c78011321d4b"} Oct 01 09:46:52 crc kubenswrapper[4787]: I1001 09:46:52.785613 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5c65595bb4-dvx6f" event={"ID":"2e57c36d-b34a-43eb-8c96-15956f1bfacd","Type":"ContainerStarted","Data":"51cf5877e51149692fb32be657f78c8d7c7d37378916d0c545861c3833eeb2f0"} Oct 01 09:46:52 crc kubenswrapper[4787]: I1001 09:46:52.814893 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5c65595bb4-dvx6f" podStartSLOduration=1.814862051 podStartE2EDuration="1.814862051s" podCreationTimestamp="2025-10-01 09:46:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:46:52.811137444 +0000 UTC m=+644.926281641" watchObservedRunningTime="2025-10-01 09:46:52.814862051 +0000 UTC m=+644.930006218" Oct 01 09:46:54 crc kubenswrapper[4787]: I1001 09:46:54.798001 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-2rzr2" event={"ID":"ee9e051d-7391-4a43-9a02-5084da577b36","Type":"ContainerStarted","Data":"2f7c4d53fdb838c80f19e82793525291f20cb022ea7018378078c29bf27dbcbb"} Oct 01 09:46:54 crc kubenswrapper[4787]: I1001 09:46:54.798431 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-2rzr2" Oct 01 09:46:54 crc kubenswrapper[4787]: I1001 09:46:54.800993 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58fcddf996-2cnst" event={"ID":"10cd7a64-357c-4c0d-b11c-426e4060b61c","Type":"ContainerStarted","Data":"212b3fc314e79032b7e049d6babf50f8039b9951a39f16d8f5f26516374ee3aa"} Oct 01 09:46:54 crc kubenswrapper[4787]: I1001 09:46:54.802463 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6d689559c5-22sp6" event={"ID":"97efd9a6-e221-4aac-97b1-b49de45edd88","Type":"ContainerStarted","Data":"777d4dc3c6b01b2d61e4b99fa6fa16627d8e55bd12f801953f15dc0ed7b2f63a"} Oct 01 09:46:54 crc kubenswrapper[4787]: I1001 09:46:54.802578 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6d689559c5-22sp6" Oct 01 09:46:54 crc kubenswrapper[4787]: I1001 09:46:54.803791 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-cjhqm" event={"ID":"e4a805b1-d488-4007-9dd3-38d3fbab1f10","Type":"ContainerStarted","Data":"a49a6225224fe1be9bc850626b4839eca431ece57e8ea19eda40b06d1b8739a2"} Oct 01 09:46:54 crc kubenswrapper[4787]: I1001 09:46:54.827798 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6d689559c5-22sp6" podStartSLOduration=1.00699366 podStartE2EDuration="3.827778228s" podCreationTimestamp="2025-10-01 09:46:51 +0000 UTC" firstStartedPulling="2025-10-01 09:46:51.662879612 +0000 UTC m=+643.778023759" lastFinishedPulling="2025-10-01 09:46:54.48366416 +0000 UTC m=+646.598808327" observedRunningTime="2025-10-01 09:46:54.827285495 +0000 UTC m=+646.942429662" watchObservedRunningTime="2025-10-01 09:46:54.827778228 +0000 UTC m=+646.942922385" Oct 01 09:46:54 crc kubenswrapper[4787]: I1001 09:46:54.829519 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-2rzr2" podStartSLOduration=0.810531494 podStartE2EDuration="3.829509043s" podCreationTimestamp="2025-10-01 09:46:51 +0000 UTC" firstStartedPulling="2025-10-01 09:46:51.457112444 +0000 UTC m=+643.572256601" lastFinishedPulling="2025-10-01 09:46:54.476089993 +0000 UTC m=+646.591234150" observedRunningTime="2025-10-01 09:46:54.812306047 +0000 UTC m=+646.927450204" watchObservedRunningTime="2025-10-01 09:46:54.829509043 +0000 UTC m=+646.944653200" Oct 01 09:46:54 crc kubenswrapper[4787]: I1001 09:46:54.851758 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-cjhqm" podStartSLOduration=1.146216504 podStartE2EDuration="3.85174124s" podCreationTimestamp="2025-10-01 09:46:51 +0000 UTC" firstStartedPulling="2025-10-01 09:46:51.761466241 +0000 UTC m=+643.876610398" lastFinishedPulling="2025-10-01 09:46:54.466990967 +0000 UTC m=+646.582135134" observedRunningTime="2025-10-01 09:46:54.848477685 +0000 UTC m=+646.963621842" watchObservedRunningTime="2025-10-01 09:46:54.85174124 +0000 UTC m=+646.966885397" Oct 01 09:46:57 crc kubenswrapper[4787]: I1001 09:46:57.823204 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58fcddf996-2cnst" event={"ID":"10cd7a64-357c-4c0d-b11c-426e4060b61c","Type":"ContainerStarted","Data":"a9b8b4a977cdf8fddb3408ae2b344962f1ecea998da6dbf3aaaa7dac1bec48d2"} Oct 01 09:46:57 crc kubenswrapper[4787]: I1001 09:46:57.840232 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-58fcddf996-2cnst" podStartSLOduration=1.835254441 podStartE2EDuration="6.840216129s" podCreationTimestamp="2025-10-01 09:46:51 +0000 UTC" firstStartedPulling="2025-10-01 09:46:51.896609167 +0000 UTC m=+644.011753324" lastFinishedPulling="2025-10-01 09:46:56.901570855 +0000 UTC m=+649.016715012" observedRunningTime="2025-10-01 09:46:57.83795791 +0000 UTC m=+649.953102077" watchObservedRunningTime="2025-10-01 09:46:57.840216129 +0000 UTC m=+649.955360286" Oct 01 09:47:01 crc kubenswrapper[4787]: I1001 09:47:01.433298 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-2rzr2" Oct 01 09:47:01 crc kubenswrapper[4787]: I1001 09:47:01.753578 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5c65595bb4-dvx6f" Oct 01 09:47:01 crc kubenswrapper[4787]: I1001 09:47:01.753627 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5c65595bb4-dvx6f" Oct 01 09:47:01 crc kubenswrapper[4787]: I1001 09:47:01.757841 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5c65595bb4-dvx6f" Oct 01 09:47:01 crc kubenswrapper[4787]: I1001 09:47:01.847590 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5c65595bb4-dvx6f" Oct 01 09:47:01 crc kubenswrapper[4787]: I1001 09:47:01.898407 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-gqvt9"] Oct 01 09:47:11 crc kubenswrapper[4787]: I1001 09:47:11.401481 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6d689559c5-22sp6" Oct 01 09:47:23 crc kubenswrapper[4787]: I1001 09:47:23.160285 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft"] Oct 01 09:47:23 crc kubenswrapper[4787]: I1001 09:47:23.161903 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft" Oct 01 09:47:23 crc kubenswrapper[4787]: I1001 09:47:23.165292 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 01 09:47:23 crc kubenswrapper[4787]: I1001 09:47:23.181839 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft"] Oct 01 09:47:23 crc kubenswrapper[4787]: I1001 09:47:23.262975 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mm7gm\" (UniqueName: \"kubernetes.io/projected/810c9efa-ff37-4f0a-8c30-d620a00bd8a4-kube-api-access-mm7gm\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft\" (UID: \"810c9efa-ff37-4f0a-8c30-d620a00bd8a4\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft" Oct 01 09:47:23 crc kubenswrapper[4787]: I1001 09:47:23.263283 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/810c9efa-ff37-4f0a-8c30-d620a00bd8a4-util\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft\" (UID: \"810c9efa-ff37-4f0a-8c30-d620a00bd8a4\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft" Oct 01 09:47:23 crc kubenswrapper[4787]: I1001 09:47:23.263413 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/810c9efa-ff37-4f0a-8c30-d620a00bd8a4-bundle\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft\" (UID: \"810c9efa-ff37-4f0a-8c30-d620a00bd8a4\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft" Oct 01 09:47:23 crc kubenswrapper[4787]: I1001 09:47:23.364371 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mm7gm\" (UniqueName: \"kubernetes.io/projected/810c9efa-ff37-4f0a-8c30-d620a00bd8a4-kube-api-access-mm7gm\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft\" (UID: \"810c9efa-ff37-4f0a-8c30-d620a00bd8a4\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft" Oct 01 09:47:23 crc kubenswrapper[4787]: I1001 09:47:23.364449 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/810c9efa-ff37-4f0a-8c30-d620a00bd8a4-util\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft\" (UID: \"810c9efa-ff37-4f0a-8c30-d620a00bd8a4\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft" Oct 01 09:47:23 crc kubenswrapper[4787]: I1001 09:47:23.364508 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/810c9efa-ff37-4f0a-8c30-d620a00bd8a4-bundle\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft\" (UID: \"810c9efa-ff37-4f0a-8c30-d620a00bd8a4\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft" Oct 01 09:47:23 crc kubenswrapper[4787]: I1001 09:47:23.365545 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/810c9efa-ff37-4f0a-8c30-d620a00bd8a4-bundle\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft\" (UID: \"810c9efa-ff37-4f0a-8c30-d620a00bd8a4\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft" Oct 01 09:47:23 crc kubenswrapper[4787]: I1001 09:47:23.365621 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/810c9efa-ff37-4f0a-8c30-d620a00bd8a4-util\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft\" (UID: \"810c9efa-ff37-4f0a-8c30-d620a00bd8a4\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft" Oct 01 09:47:23 crc kubenswrapper[4787]: I1001 09:47:23.385851 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mm7gm\" (UniqueName: \"kubernetes.io/projected/810c9efa-ff37-4f0a-8c30-d620a00bd8a4-kube-api-access-mm7gm\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft\" (UID: \"810c9efa-ff37-4f0a-8c30-d620a00bd8a4\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft" Oct 01 09:47:23 crc kubenswrapper[4787]: I1001 09:47:23.477792 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft" Oct 01 09:47:23 crc kubenswrapper[4787]: I1001 09:47:23.676194 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft"] Oct 01 09:47:23 crc kubenswrapper[4787]: I1001 09:47:23.962118 4787 generic.go:334] "Generic (PLEG): container finished" podID="810c9efa-ff37-4f0a-8c30-d620a00bd8a4" containerID="a0b05428f6f47dc3f3bb5e4f220aa2f61b42ade9591685b5ae647f7caafe4d00" exitCode=0 Oct 01 09:47:23 crc kubenswrapper[4787]: I1001 09:47:23.962184 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft" event={"ID":"810c9efa-ff37-4f0a-8c30-d620a00bd8a4","Type":"ContainerDied","Data":"a0b05428f6f47dc3f3bb5e4f220aa2f61b42ade9591685b5ae647f7caafe4d00"} Oct 01 09:47:23 crc kubenswrapper[4787]: I1001 09:47:23.962221 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft" event={"ID":"810c9efa-ff37-4f0a-8c30-d620a00bd8a4","Type":"ContainerStarted","Data":"d7c7977a2866a15a10707d7fd574f30026507d5d4b4a26865154fbf05afacb48"} Oct 01 09:47:25 crc kubenswrapper[4787]: I1001 09:47:25.996889 4787 generic.go:334] "Generic (PLEG): container finished" podID="810c9efa-ff37-4f0a-8c30-d620a00bd8a4" containerID="32396aefee772533d6190ec72b235b2c0d547a1ada62425bfaba49276b38d15e" exitCode=0 Oct 01 09:47:25 crc kubenswrapper[4787]: I1001 09:47:25.997041 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft" event={"ID":"810c9efa-ff37-4f0a-8c30-d620a00bd8a4","Type":"ContainerDied","Data":"32396aefee772533d6190ec72b235b2c0d547a1ada62425bfaba49276b38d15e"} Oct 01 09:47:26 crc kubenswrapper[4787]: I1001 09:47:26.944459 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-gqvt9" podUID="2ec1df2f-eadd-4385-97b4-664101fb9d51" containerName="console" containerID="cri-o://a4c271522ac1c586ceb37ed14cec52855bc91429be6335c564b8a0318fb00173" gracePeriod=15 Oct 01 09:47:27 crc kubenswrapper[4787]: I1001 09:47:27.007004 4787 generic.go:334] "Generic (PLEG): container finished" podID="810c9efa-ff37-4f0a-8c30-d620a00bd8a4" containerID="741e6bd2acd0225e653814a7ea5616d1df4d69eec1e5db34a6267b6bcd1466fd" exitCode=0 Oct 01 09:47:27 crc kubenswrapper[4787]: I1001 09:47:27.007175 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft" event={"ID":"810c9efa-ff37-4f0a-8c30-d620a00bd8a4","Type":"ContainerDied","Data":"741e6bd2acd0225e653814a7ea5616d1df4d69eec1e5db34a6267b6bcd1466fd"} Oct 01 09:47:28 crc kubenswrapper[4787]: I1001 09:47:28.016885 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-gqvt9_2ec1df2f-eadd-4385-97b4-664101fb9d51/console/0.log" Oct 01 09:47:28 crc kubenswrapper[4787]: I1001 09:47:28.017492 4787 generic.go:334] "Generic (PLEG): container finished" podID="2ec1df2f-eadd-4385-97b4-664101fb9d51" containerID="a4c271522ac1c586ceb37ed14cec52855bc91429be6335c564b8a0318fb00173" exitCode=2 Oct 01 09:47:28 crc kubenswrapper[4787]: I1001 09:47:28.017632 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-gqvt9" event={"ID":"2ec1df2f-eadd-4385-97b4-664101fb9d51","Type":"ContainerDied","Data":"a4c271522ac1c586ceb37ed14cec52855bc91429be6335c564b8a0318fb00173"} Oct 01 09:47:28 crc kubenswrapper[4787]: I1001 09:47:28.214550 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-gqvt9_2ec1df2f-eadd-4385-97b4-664101fb9d51/console/0.log" Oct 01 09:47:28 crc kubenswrapper[4787]: I1001 09:47:28.214616 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-gqvt9" Oct 01 09:47:28 crc kubenswrapper[4787]: I1001 09:47:28.284318 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft" Oct 01 09:47:28 crc kubenswrapper[4787]: I1001 09:47:28.330894 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2ec1df2f-eadd-4385-97b4-664101fb9d51-service-ca\") pod \"2ec1df2f-eadd-4385-97b4-664101fb9d51\" (UID: \"2ec1df2f-eadd-4385-97b4-664101fb9d51\") " Oct 01 09:47:28 crc kubenswrapper[4787]: I1001 09:47:28.330946 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2ec1df2f-eadd-4385-97b4-664101fb9d51-trusted-ca-bundle\") pod \"2ec1df2f-eadd-4385-97b4-664101fb9d51\" (UID: \"2ec1df2f-eadd-4385-97b4-664101fb9d51\") " Oct 01 09:47:28 crc kubenswrapper[4787]: I1001 09:47:28.330979 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2ec1df2f-eadd-4385-97b4-664101fb9d51-console-serving-cert\") pod \"2ec1df2f-eadd-4385-97b4-664101fb9d51\" (UID: \"2ec1df2f-eadd-4385-97b4-664101fb9d51\") " Oct 01 09:47:28 crc kubenswrapper[4787]: I1001 09:47:28.331010 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2ec1df2f-eadd-4385-97b4-664101fb9d51-console-oauth-config\") pod \"2ec1df2f-eadd-4385-97b4-664101fb9d51\" (UID: \"2ec1df2f-eadd-4385-97b4-664101fb9d51\") " Oct 01 09:47:28 crc kubenswrapper[4787]: I1001 09:47:28.331068 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2ec1df2f-eadd-4385-97b4-664101fb9d51-console-config\") pod \"2ec1df2f-eadd-4385-97b4-664101fb9d51\" (UID: \"2ec1df2f-eadd-4385-97b4-664101fb9d51\") " Oct 01 09:47:28 crc kubenswrapper[4787]: I1001 09:47:28.331283 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2ec1df2f-eadd-4385-97b4-664101fb9d51-oauth-serving-cert\") pod \"2ec1df2f-eadd-4385-97b4-664101fb9d51\" (UID: \"2ec1df2f-eadd-4385-97b4-664101fb9d51\") " Oct 01 09:47:28 crc kubenswrapper[4787]: I1001 09:47:28.331317 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7f2hj\" (UniqueName: \"kubernetes.io/projected/2ec1df2f-eadd-4385-97b4-664101fb9d51-kube-api-access-7f2hj\") pod \"2ec1df2f-eadd-4385-97b4-664101fb9d51\" (UID: \"2ec1df2f-eadd-4385-97b4-664101fb9d51\") " Oct 01 09:47:28 crc kubenswrapper[4787]: I1001 09:47:28.332051 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ec1df2f-eadd-4385-97b4-664101fb9d51-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "2ec1df2f-eadd-4385-97b4-664101fb9d51" (UID: "2ec1df2f-eadd-4385-97b4-664101fb9d51"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:47:28 crc kubenswrapper[4787]: I1001 09:47:28.332197 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ec1df2f-eadd-4385-97b4-664101fb9d51-service-ca" (OuterVolumeSpecName: "service-ca") pod "2ec1df2f-eadd-4385-97b4-664101fb9d51" (UID: "2ec1df2f-eadd-4385-97b4-664101fb9d51"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:47:28 crc kubenswrapper[4787]: I1001 09:47:28.333018 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ec1df2f-eadd-4385-97b4-664101fb9d51-console-config" (OuterVolumeSpecName: "console-config") pod "2ec1df2f-eadd-4385-97b4-664101fb9d51" (UID: "2ec1df2f-eadd-4385-97b4-664101fb9d51"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:47:28 crc kubenswrapper[4787]: I1001 09:47:28.333029 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ec1df2f-eadd-4385-97b4-664101fb9d51-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "2ec1df2f-eadd-4385-97b4-664101fb9d51" (UID: "2ec1df2f-eadd-4385-97b4-664101fb9d51"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:47:28 crc kubenswrapper[4787]: I1001 09:47:28.337209 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ec1df2f-eadd-4385-97b4-664101fb9d51-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "2ec1df2f-eadd-4385-97b4-664101fb9d51" (UID: "2ec1df2f-eadd-4385-97b4-664101fb9d51"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:47:28 crc kubenswrapper[4787]: I1001 09:47:28.337440 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ec1df2f-eadd-4385-97b4-664101fb9d51-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "2ec1df2f-eadd-4385-97b4-664101fb9d51" (UID: "2ec1df2f-eadd-4385-97b4-664101fb9d51"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:47:28 crc kubenswrapper[4787]: I1001 09:47:28.341115 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ec1df2f-eadd-4385-97b4-664101fb9d51-kube-api-access-7f2hj" (OuterVolumeSpecName: "kube-api-access-7f2hj") pod "2ec1df2f-eadd-4385-97b4-664101fb9d51" (UID: "2ec1df2f-eadd-4385-97b4-664101fb9d51"). InnerVolumeSpecName "kube-api-access-7f2hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:47:28 crc kubenswrapper[4787]: I1001 09:47:28.432497 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/810c9efa-ff37-4f0a-8c30-d620a00bd8a4-bundle\") pod \"810c9efa-ff37-4f0a-8c30-d620a00bd8a4\" (UID: \"810c9efa-ff37-4f0a-8c30-d620a00bd8a4\") " Oct 01 09:47:28 crc kubenswrapper[4787]: I1001 09:47:28.432590 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/810c9efa-ff37-4f0a-8c30-d620a00bd8a4-util\") pod \"810c9efa-ff37-4f0a-8c30-d620a00bd8a4\" (UID: \"810c9efa-ff37-4f0a-8c30-d620a00bd8a4\") " Oct 01 09:47:28 crc kubenswrapper[4787]: I1001 09:47:28.432630 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mm7gm\" (UniqueName: \"kubernetes.io/projected/810c9efa-ff37-4f0a-8c30-d620a00bd8a4-kube-api-access-mm7gm\") pod \"810c9efa-ff37-4f0a-8c30-d620a00bd8a4\" (UID: \"810c9efa-ff37-4f0a-8c30-d620a00bd8a4\") " Oct 01 09:47:28 crc kubenswrapper[4787]: I1001 09:47:28.433021 4787 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2ec1df2f-eadd-4385-97b4-664101fb9d51-service-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:47:28 crc kubenswrapper[4787]: I1001 09:47:28.433051 4787 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2ec1df2f-eadd-4385-97b4-664101fb9d51-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:47:28 crc kubenswrapper[4787]: I1001 09:47:28.433066 4787 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2ec1df2f-eadd-4385-97b4-664101fb9d51-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:47:28 crc kubenswrapper[4787]: I1001 09:47:28.433095 4787 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2ec1df2f-eadd-4385-97b4-664101fb9d51-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:47:28 crc kubenswrapper[4787]: I1001 09:47:28.433107 4787 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2ec1df2f-eadd-4385-97b4-664101fb9d51-console-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:47:28 crc kubenswrapper[4787]: I1001 09:47:28.433121 4787 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2ec1df2f-eadd-4385-97b4-664101fb9d51-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:47:28 crc kubenswrapper[4787]: I1001 09:47:28.433132 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7f2hj\" (UniqueName: \"kubernetes.io/projected/2ec1df2f-eadd-4385-97b4-664101fb9d51-kube-api-access-7f2hj\") on node \"crc\" DevicePath \"\"" Oct 01 09:47:28 crc kubenswrapper[4787]: I1001 09:47:28.433848 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/810c9efa-ff37-4f0a-8c30-d620a00bd8a4-bundle" (OuterVolumeSpecName: "bundle") pod "810c9efa-ff37-4f0a-8c30-d620a00bd8a4" (UID: "810c9efa-ff37-4f0a-8c30-d620a00bd8a4"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:47:28 crc kubenswrapper[4787]: I1001 09:47:28.435797 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/810c9efa-ff37-4f0a-8c30-d620a00bd8a4-kube-api-access-mm7gm" (OuterVolumeSpecName: "kube-api-access-mm7gm") pod "810c9efa-ff37-4f0a-8c30-d620a00bd8a4" (UID: "810c9efa-ff37-4f0a-8c30-d620a00bd8a4"). InnerVolumeSpecName "kube-api-access-mm7gm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:47:28 crc kubenswrapper[4787]: I1001 09:47:28.448229 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/810c9efa-ff37-4f0a-8c30-d620a00bd8a4-util" (OuterVolumeSpecName: "util") pod "810c9efa-ff37-4f0a-8c30-d620a00bd8a4" (UID: "810c9efa-ff37-4f0a-8c30-d620a00bd8a4"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:47:28 crc kubenswrapper[4787]: I1001 09:47:28.535136 4787 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/810c9efa-ff37-4f0a-8c30-d620a00bd8a4-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:47:28 crc kubenswrapper[4787]: I1001 09:47:28.535201 4787 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/810c9efa-ff37-4f0a-8c30-d620a00bd8a4-util\") on node \"crc\" DevicePath \"\"" Oct 01 09:47:28 crc kubenswrapper[4787]: I1001 09:47:28.535222 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mm7gm\" (UniqueName: \"kubernetes.io/projected/810c9efa-ff37-4f0a-8c30-d620a00bd8a4-kube-api-access-mm7gm\") on node \"crc\" DevicePath \"\"" Oct 01 09:47:29 crc kubenswrapper[4787]: I1001 09:47:29.026003 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft" event={"ID":"810c9efa-ff37-4f0a-8c30-d620a00bd8a4","Type":"ContainerDied","Data":"d7c7977a2866a15a10707d7fd574f30026507d5d4b4a26865154fbf05afacb48"} Oct 01 09:47:29 crc kubenswrapper[4787]: I1001 09:47:29.026032 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft" Oct 01 09:47:29 crc kubenswrapper[4787]: I1001 09:47:29.026048 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7c7977a2866a15a10707d7fd574f30026507d5d4b4a26865154fbf05afacb48" Oct 01 09:47:29 crc kubenswrapper[4787]: I1001 09:47:29.029107 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-gqvt9_2ec1df2f-eadd-4385-97b4-664101fb9d51/console/0.log" Oct 01 09:47:29 crc kubenswrapper[4787]: I1001 09:47:29.029352 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-gqvt9" event={"ID":"2ec1df2f-eadd-4385-97b4-664101fb9d51","Type":"ContainerDied","Data":"7fb44713d867d5f5b7e61272f9947cc5ddbb9c637b20236ca48d6ba01ac38bb3"} Oct 01 09:47:29 crc kubenswrapper[4787]: I1001 09:47:29.029443 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-gqvt9" Oct 01 09:47:29 crc kubenswrapper[4787]: I1001 09:47:29.029534 4787 scope.go:117] "RemoveContainer" containerID="a4c271522ac1c586ceb37ed14cec52855bc91429be6335c564b8a0318fb00173" Oct 01 09:47:29 crc kubenswrapper[4787]: I1001 09:47:29.072023 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-gqvt9"] Oct 01 09:47:29 crc kubenswrapper[4787]: I1001 09:47:29.078248 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-gqvt9"] Oct 01 09:47:30 crc kubenswrapper[4787]: I1001 09:47:30.533668 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ec1df2f-eadd-4385-97b4-664101fb9d51" path="/var/lib/kubelet/pods/2ec1df2f-eadd-4385-97b4-664101fb9d51/volumes" Oct 01 09:47:37 crc kubenswrapper[4787]: I1001 09:47:37.739898 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-86f7dcd4cf-c4xjr"] Oct 01 09:47:37 crc kubenswrapper[4787]: E1001 09:47:37.740804 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ec1df2f-eadd-4385-97b4-664101fb9d51" containerName="console" Oct 01 09:47:37 crc kubenswrapper[4787]: I1001 09:47:37.740821 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ec1df2f-eadd-4385-97b4-664101fb9d51" containerName="console" Oct 01 09:47:37 crc kubenswrapper[4787]: E1001 09:47:37.740839 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="810c9efa-ff37-4f0a-8c30-d620a00bd8a4" containerName="util" Oct 01 09:47:37 crc kubenswrapper[4787]: I1001 09:47:37.740847 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="810c9efa-ff37-4f0a-8c30-d620a00bd8a4" containerName="util" Oct 01 09:47:37 crc kubenswrapper[4787]: E1001 09:47:37.740864 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="810c9efa-ff37-4f0a-8c30-d620a00bd8a4" containerName="pull" Oct 01 09:47:37 crc kubenswrapper[4787]: I1001 09:47:37.740873 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="810c9efa-ff37-4f0a-8c30-d620a00bd8a4" containerName="pull" Oct 01 09:47:37 crc kubenswrapper[4787]: E1001 09:47:37.740885 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="810c9efa-ff37-4f0a-8c30-d620a00bd8a4" containerName="extract" Oct 01 09:47:37 crc kubenswrapper[4787]: I1001 09:47:37.740892 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="810c9efa-ff37-4f0a-8c30-d620a00bd8a4" containerName="extract" Oct 01 09:47:37 crc kubenswrapper[4787]: I1001 09:47:37.741059 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ec1df2f-eadd-4385-97b4-664101fb9d51" containerName="console" Oct 01 09:47:37 crc kubenswrapper[4787]: I1001 09:47:37.741091 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="810c9efa-ff37-4f0a-8c30-d620a00bd8a4" containerName="extract" Oct 01 09:47:37 crc kubenswrapper[4787]: I1001 09:47:37.741650 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-86f7dcd4cf-c4xjr" Oct 01 09:47:37 crc kubenswrapper[4787]: I1001 09:47:37.744223 4787 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 01 09:47:37 crc kubenswrapper[4787]: I1001 09:47:37.744223 4787 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 01 09:47:37 crc kubenswrapper[4787]: I1001 09:47:37.746035 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 01 09:47:37 crc kubenswrapper[4787]: I1001 09:47:37.746184 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 01 09:47:37 crc kubenswrapper[4787]: I1001 09:47:37.747577 4787 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-gpjgs" Oct 01 09:47:37 crc kubenswrapper[4787]: I1001 09:47:37.759065 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4l8dp\" (UniqueName: \"kubernetes.io/projected/178c8464-8b87-4dc1-9c52-7dc1f5b3714d-kube-api-access-4l8dp\") pod \"metallb-operator-controller-manager-86f7dcd4cf-c4xjr\" (UID: \"178c8464-8b87-4dc1-9c52-7dc1f5b3714d\") " pod="metallb-system/metallb-operator-controller-manager-86f7dcd4cf-c4xjr" Oct 01 09:47:37 crc kubenswrapper[4787]: I1001 09:47:37.759163 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/178c8464-8b87-4dc1-9c52-7dc1f5b3714d-apiservice-cert\") pod \"metallb-operator-controller-manager-86f7dcd4cf-c4xjr\" (UID: \"178c8464-8b87-4dc1-9c52-7dc1f5b3714d\") " pod="metallb-system/metallb-operator-controller-manager-86f7dcd4cf-c4xjr" Oct 01 09:47:37 crc kubenswrapper[4787]: I1001 09:47:37.759214 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/178c8464-8b87-4dc1-9c52-7dc1f5b3714d-webhook-cert\") pod \"metallb-operator-controller-manager-86f7dcd4cf-c4xjr\" (UID: \"178c8464-8b87-4dc1-9c52-7dc1f5b3714d\") " pod="metallb-system/metallb-operator-controller-manager-86f7dcd4cf-c4xjr" Oct 01 09:47:37 crc kubenswrapper[4787]: I1001 09:47:37.763371 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-86f7dcd4cf-c4xjr"] Oct 01 09:47:37 crc kubenswrapper[4787]: I1001 09:47:37.860637 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/178c8464-8b87-4dc1-9c52-7dc1f5b3714d-webhook-cert\") pod \"metallb-operator-controller-manager-86f7dcd4cf-c4xjr\" (UID: \"178c8464-8b87-4dc1-9c52-7dc1f5b3714d\") " pod="metallb-system/metallb-operator-controller-manager-86f7dcd4cf-c4xjr" Oct 01 09:47:37 crc kubenswrapper[4787]: I1001 09:47:37.860727 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4l8dp\" (UniqueName: \"kubernetes.io/projected/178c8464-8b87-4dc1-9c52-7dc1f5b3714d-kube-api-access-4l8dp\") pod \"metallb-operator-controller-manager-86f7dcd4cf-c4xjr\" (UID: \"178c8464-8b87-4dc1-9c52-7dc1f5b3714d\") " pod="metallb-system/metallb-operator-controller-manager-86f7dcd4cf-c4xjr" Oct 01 09:47:37 crc kubenswrapper[4787]: I1001 09:47:37.860754 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/178c8464-8b87-4dc1-9c52-7dc1f5b3714d-apiservice-cert\") pod \"metallb-operator-controller-manager-86f7dcd4cf-c4xjr\" (UID: \"178c8464-8b87-4dc1-9c52-7dc1f5b3714d\") " pod="metallb-system/metallb-operator-controller-manager-86f7dcd4cf-c4xjr" Oct 01 09:47:37 crc kubenswrapper[4787]: I1001 09:47:37.866956 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/178c8464-8b87-4dc1-9c52-7dc1f5b3714d-apiservice-cert\") pod \"metallb-operator-controller-manager-86f7dcd4cf-c4xjr\" (UID: \"178c8464-8b87-4dc1-9c52-7dc1f5b3714d\") " pod="metallb-system/metallb-operator-controller-manager-86f7dcd4cf-c4xjr" Oct 01 09:47:37 crc kubenswrapper[4787]: I1001 09:47:37.866962 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/178c8464-8b87-4dc1-9c52-7dc1f5b3714d-webhook-cert\") pod \"metallb-operator-controller-manager-86f7dcd4cf-c4xjr\" (UID: \"178c8464-8b87-4dc1-9c52-7dc1f5b3714d\") " pod="metallb-system/metallb-operator-controller-manager-86f7dcd4cf-c4xjr" Oct 01 09:47:37 crc kubenswrapper[4787]: I1001 09:47:37.880433 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4l8dp\" (UniqueName: \"kubernetes.io/projected/178c8464-8b87-4dc1-9c52-7dc1f5b3714d-kube-api-access-4l8dp\") pod \"metallb-operator-controller-manager-86f7dcd4cf-c4xjr\" (UID: \"178c8464-8b87-4dc1-9c52-7dc1f5b3714d\") " pod="metallb-system/metallb-operator-controller-manager-86f7dcd4cf-c4xjr" Oct 01 09:47:38 crc kubenswrapper[4787]: I1001 09:47:38.058554 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-86f7dcd4cf-c4xjr" Oct 01 09:47:38 crc kubenswrapper[4787]: I1001 09:47:38.060917 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-7d48896987-v5r98"] Oct 01 09:47:38 crc kubenswrapper[4787]: I1001 09:47:38.061776 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7d48896987-v5r98" Oct 01 09:47:38 crc kubenswrapper[4787]: I1001 09:47:38.064093 4787 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-zcslf" Oct 01 09:47:38 crc kubenswrapper[4787]: I1001 09:47:38.064103 4787 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 01 09:47:38 crc kubenswrapper[4787]: I1001 09:47:38.064147 4787 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 01 09:47:38 crc kubenswrapper[4787]: I1001 09:47:38.083351 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7d48896987-v5r98"] Oct 01 09:47:38 crc kubenswrapper[4787]: I1001 09:47:38.163911 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnq9n\" (UniqueName: \"kubernetes.io/projected/55c631b4-c7e0-47cb-9e66-93b8a8c51877-kube-api-access-qnq9n\") pod \"metallb-operator-webhook-server-7d48896987-v5r98\" (UID: \"55c631b4-c7e0-47cb-9e66-93b8a8c51877\") " pod="metallb-system/metallb-operator-webhook-server-7d48896987-v5r98" Oct 01 09:47:38 crc kubenswrapper[4787]: I1001 09:47:38.164002 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/55c631b4-c7e0-47cb-9e66-93b8a8c51877-apiservice-cert\") pod \"metallb-operator-webhook-server-7d48896987-v5r98\" (UID: \"55c631b4-c7e0-47cb-9e66-93b8a8c51877\") " pod="metallb-system/metallb-operator-webhook-server-7d48896987-v5r98" Oct 01 09:47:38 crc kubenswrapper[4787]: I1001 09:47:38.164063 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/55c631b4-c7e0-47cb-9e66-93b8a8c51877-webhook-cert\") pod \"metallb-operator-webhook-server-7d48896987-v5r98\" (UID: \"55c631b4-c7e0-47cb-9e66-93b8a8c51877\") " pod="metallb-system/metallb-operator-webhook-server-7d48896987-v5r98" Oct 01 09:47:38 crc kubenswrapper[4787]: I1001 09:47:38.264954 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnq9n\" (UniqueName: \"kubernetes.io/projected/55c631b4-c7e0-47cb-9e66-93b8a8c51877-kube-api-access-qnq9n\") pod \"metallb-operator-webhook-server-7d48896987-v5r98\" (UID: \"55c631b4-c7e0-47cb-9e66-93b8a8c51877\") " pod="metallb-system/metallb-operator-webhook-server-7d48896987-v5r98" Oct 01 09:47:38 crc kubenswrapper[4787]: I1001 09:47:38.265509 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/55c631b4-c7e0-47cb-9e66-93b8a8c51877-apiservice-cert\") pod \"metallb-operator-webhook-server-7d48896987-v5r98\" (UID: \"55c631b4-c7e0-47cb-9e66-93b8a8c51877\") " pod="metallb-system/metallb-operator-webhook-server-7d48896987-v5r98" Oct 01 09:47:38 crc kubenswrapper[4787]: I1001 09:47:38.265554 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/55c631b4-c7e0-47cb-9e66-93b8a8c51877-webhook-cert\") pod \"metallb-operator-webhook-server-7d48896987-v5r98\" (UID: \"55c631b4-c7e0-47cb-9e66-93b8a8c51877\") " pod="metallb-system/metallb-operator-webhook-server-7d48896987-v5r98" Oct 01 09:47:38 crc kubenswrapper[4787]: I1001 09:47:38.272275 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/55c631b4-c7e0-47cb-9e66-93b8a8c51877-webhook-cert\") pod \"metallb-operator-webhook-server-7d48896987-v5r98\" (UID: \"55c631b4-c7e0-47cb-9e66-93b8a8c51877\") " pod="metallb-system/metallb-operator-webhook-server-7d48896987-v5r98" Oct 01 09:47:38 crc kubenswrapper[4787]: I1001 09:47:38.285808 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnq9n\" (UniqueName: \"kubernetes.io/projected/55c631b4-c7e0-47cb-9e66-93b8a8c51877-kube-api-access-qnq9n\") pod \"metallb-operator-webhook-server-7d48896987-v5r98\" (UID: \"55c631b4-c7e0-47cb-9e66-93b8a8c51877\") " pod="metallb-system/metallb-operator-webhook-server-7d48896987-v5r98" Oct 01 09:47:38 crc kubenswrapper[4787]: I1001 09:47:38.287085 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-86f7dcd4cf-c4xjr"] Oct 01 09:47:38 crc kubenswrapper[4787]: I1001 09:47:38.296585 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/55c631b4-c7e0-47cb-9e66-93b8a8c51877-apiservice-cert\") pod \"metallb-operator-webhook-server-7d48896987-v5r98\" (UID: \"55c631b4-c7e0-47cb-9e66-93b8a8c51877\") " pod="metallb-system/metallb-operator-webhook-server-7d48896987-v5r98" Oct 01 09:47:38 crc kubenswrapper[4787]: I1001 09:47:38.419286 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7d48896987-v5r98" Oct 01 09:47:38 crc kubenswrapper[4787]: W1001 09:47:38.848616 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod55c631b4_c7e0_47cb_9e66_93b8a8c51877.slice/crio-9e278ded541c46d051335b9fdb74cb8094712d94b1bd145676ea565c61240201 WatchSource:0}: Error finding container 9e278ded541c46d051335b9fdb74cb8094712d94b1bd145676ea565c61240201: Status 404 returned error can't find the container with id 9e278ded541c46d051335b9fdb74cb8094712d94b1bd145676ea565c61240201 Oct 01 09:47:38 crc kubenswrapper[4787]: I1001 09:47:38.858090 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7d48896987-v5r98"] Oct 01 09:47:39 crc kubenswrapper[4787]: I1001 09:47:39.089301 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7d48896987-v5r98" event={"ID":"55c631b4-c7e0-47cb-9e66-93b8a8c51877","Type":"ContainerStarted","Data":"9e278ded541c46d051335b9fdb74cb8094712d94b1bd145676ea565c61240201"} Oct 01 09:47:39 crc kubenswrapper[4787]: I1001 09:47:39.090473 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-86f7dcd4cf-c4xjr" event={"ID":"178c8464-8b87-4dc1-9c52-7dc1f5b3714d","Type":"ContainerStarted","Data":"702d7db242a04abf3b8cebb10fddd971d8a667a5974cf8fca11854bb71a51a88"} Oct 01 09:47:41 crc kubenswrapper[4787]: I1001 09:47:41.252030 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:47:41 crc kubenswrapper[4787]: I1001 09:47:41.252659 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:47:42 crc kubenswrapper[4787]: I1001 09:47:42.114713 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-86f7dcd4cf-c4xjr" event={"ID":"178c8464-8b87-4dc1-9c52-7dc1f5b3714d","Type":"ContainerStarted","Data":"d9d69594e88edf98f300314d7b19d45a464fe342b6a93b18155cd7290651d5a1"} Oct 01 09:47:42 crc kubenswrapper[4787]: I1001 09:47:42.115319 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-86f7dcd4cf-c4xjr" Oct 01 09:47:42 crc kubenswrapper[4787]: I1001 09:47:42.144036 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-86f7dcd4cf-c4xjr" podStartSLOduration=2.291663273 podStartE2EDuration="5.144017856s" podCreationTimestamp="2025-10-01 09:47:37 +0000 UTC" firstStartedPulling="2025-10-01 09:47:38.298417338 +0000 UTC m=+690.413561495" lastFinishedPulling="2025-10-01 09:47:41.150771891 +0000 UTC m=+693.265916078" observedRunningTime="2025-10-01 09:47:42.141657276 +0000 UTC m=+694.256801443" watchObservedRunningTime="2025-10-01 09:47:42.144017856 +0000 UTC m=+694.259162013" Oct 01 09:47:44 crc kubenswrapper[4787]: I1001 09:47:44.130828 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7d48896987-v5r98" event={"ID":"55c631b4-c7e0-47cb-9e66-93b8a8c51877","Type":"ContainerStarted","Data":"f4eac3cc96c5db09b5f5e8b9b2c90e26858d63c259fdc220e9168ea26cd52cbe"} Oct 01 09:47:44 crc kubenswrapper[4787]: I1001 09:47:44.131490 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-7d48896987-v5r98" Oct 01 09:47:44 crc kubenswrapper[4787]: I1001 09:47:44.158328 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-7d48896987-v5r98" podStartSLOduration=1.683226935 podStartE2EDuration="6.158303572s" podCreationTimestamp="2025-10-01 09:47:38 +0000 UTC" firstStartedPulling="2025-10-01 09:47:38.857771462 +0000 UTC m=+690.972915619" lastFinishedPulling="2025-10-01 09:47:43.332848099 +0000 UTC m=+695.447992256" observedRunningTime="2025-10-01 09:47:44.156128547 +0000 UTC m=+696.271272704" watchObservedRunningTime="2025-10-01 09:47:44.158303572 +0000 UTC m=+696.273447729" Oct 01 09:47:58 crc kubenswrapper[4787]: I1001 09:47:58.424778 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-7d48896987-v5r98" Oct 01 09:48:11 crc kubenswrapper[4787]: I1001 09:48:11.251161 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:48:11 crc kubenswrapper[4787]: I1001 09:48:11.252134 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:48:18 crc kubenswrapper[4787]: I1001 09:48:18.061795 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-86f7dcd4cf-c4xjr" Oct 01 09:48:18 crc kubenswrapper[4787]: I1001 09:48:18.877305 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-kcxb8"] Oct 01 09:48:18 crc kubenswrapper[4787]: I1001 09:48:18.879229 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-kcxb8" Oct 01 09:48:18 crc kubenswrapper[4787]: I1001 09:48:18.880333 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-csxhg"] Oct 01 09:48:18 crc kubenswrapper[4787]: I1001 09:48:18.883447 4787 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 01 09:48:18 crc kubenswrapper[4787]: I1001 09:48:18.883502 4787 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-mclrt" Oct 01 09:48:18 crc kubenswrapper[4787]: I1001 09:48:18.896360 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-csxhg" Oct 01 09:48:18 crc kubenswrapper[4787]: I1001 09:48:18.898924 4787 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 01 09:48:18 crc kubenswrapper[4787]: I1001 09:48:18.903571 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 01 09:48:18 crc kubenswrapper[4787]: I1001 09:48:18.939794 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-kcxb8"] Oct 01 09:48:18 crc kubenswrapper[4787]: I1001 09:48:18.969676 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwfqg\" (UniqueName: \"kubernetes.io/projected/8b191fd2-3066-4e07-9641-98e44be362f5-kube-api-access-bwfqg\") pod \"frr-k8s-webhook-server-5478bdb765-kcxb8\" (UID: \"8b191fd2-3066-4e07-9641-98e44be362f5\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-kcxb8" Oct 01 09:48:18 crc kubenswrapper[4787]: I1001 09:48:18.969821 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8b191fd2-3066-4e07-9641-98e44be362f5-cert\") pod \"frr-k8s-webhook-server-5478bdb765-kcxb8\" (UID: \"8b191fd2-3066-4e07-9641-98e44be362f5\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-kcxb8" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.032739 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-cc45k"] Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.033992 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-cc45k" Oct 01 09:48:19 crc kubenswrapper[4787]: W1001 09:48:19.042129 4787 reflector.go:561] object-"metallb-system"/"metallb-memberlist": failed to list *v1.Secret: secrets "metallb-memberlist" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Oct 01 09:48:19 crc kubenswrapper[4787]: E1001 09:48:19.042220 4787 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"metallb-memberlist\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"metallb-memberlist\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 01 09:48:19 crc kubenswrapper[4787]: W1001 09:48:19.042339 4787 reflector.go:561] object-"metallb-system"/"metallb-excludel2": failed to list *v1.ConfigMap: configmaps "metallb-excludel2" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Oct 01 09:48:19 crc kubenswrapper[4787]: E1001 09:48:19.042397 4787 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"metallb-excludel2\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"metallb-excludel2\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.050346 4787 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-7zjbc" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.050669 4787 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.071088 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/c6570c99-4cff-4194-9744-0efc00f88332-metrics\") pod \"frr-k8s-csxhg\" (UID: \"c6570c99-4cff-4194-9744-0efc00f88332\") " pod="metallb-system/frr-k8s-csxhg" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.071203 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/c6570c99-4cff-4194-9744-0efc00f88332-frr-sockets\") pod \"frr-k8s-csxhg\" (UID: \"c6570c99-4cff-4194-9744-0efc00f88332\") " pod="metallb-system/frr-k8s-csxhg" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.071257 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vj6p6\" (UniqueName: \"kubernetes.io/projected/c6570c99-4cff-4194-9744-0efc00f88332-kube-api-access-vj6p6\") pod \"frr-k8s-csxhg\" (UID: \"c6570c99-4cff-4194-9744-0efc00f88332\") " pod="metallb-system/frr-k8s-csxhg" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.071324 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c6570c99-4cff-4194-9744-0efc00f88332-metrics-certs\") pod \"frr-k8s-csxhg\" (UID: \"c6570c99-4cff-4194-9744-0efc00f88332\") " pod="metallb-system/frr-k8s-csxhg" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.071363 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/c6570c99-4cff-4194-9744-0efc00f88332-frr-conf\") pod \"frr-k8s-csxhg\" (UID: \"c6570c99-4cff-4194-9744-0efc00f88332\") " pod="metallb-system/frr-k8s-csxhg" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.071389 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwfqg\" (UniqueName: \"kubernetes.io/projected/8b191fd2-3066-4e07-9641-98e44be362f5-kube-api-access-bwfqg\") pod \"frr-k8s-webhook-server-5478bdb765-kcxb8\" (UID: \"8b191fd2-3066-4e07-9641-98e44be362f5\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-kcxb8" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.071509 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8b191fd2-3066-4e07-9641-98e44be362f5-cert\") pod \"frr-k8s-webhook-server-5478bdb765-kcxb8\" (UID: \"8b191fd2-3066-4e07-9641-98e44be362f5\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-kcxb8" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.071559 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/c6570c99-4cff-4194-9744-0efc00f88332-frr-startup\") pod \"frr-k8s-csxhg\" (UID: \"c6570c99-4cff-4194-9744-0efc00f88332\") " pod="metallb-system/frr-k8s-csxhg" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.071613 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/c6570c99-4cff-4194-9744-0efc00f88332-reloader\") pod \"frr-k8s-csxhg\" (UID: \"c6570c99-4cff-4194-9744-0efc00f88332\") " pod="metallb-system/frr-k8s-csxhg" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.076602 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-5d688f5ffc-gnclz"] Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.077726 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5d688f5ffc-gnclz" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.091519 4787 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.092260 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8b191fd2-3066-4e07-9641-98e44be362f5-cert\") pod \"frr-k8s-webhook-server-5478bdb765-kcxb8\" (UID: \"8b191fd2-3066-4e07-9641-98e44be362f5\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-kcxb8" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.094671 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwfqg\" (UniqueName: \"kubernetes.io/projected/8b191fd2-3066-4e07-9641-98e44be362f5-kube-api-access-bwfqg\") pod \"frr-k8s-webhook-server-5478bdb765-kcxb8\" (UID: \"8b191fd2-3066-4e07-9641-98e44be362f5\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-kcxb8" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.099857 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5d688f5ffc-gnclz"] Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.172964 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c6570c99-4cff-4194-9744-0efc00f88332-metrics-certs\") pod \"frr-k8s-csxhg\" (UID: \"c6570c99-4cff-4194-9744-0efc00f88332\") " pod="metallb-system/frr-k8s-csxhg" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.173033 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668-metrics-certs\") pod \"speaker-cc45k\" (UID: \"c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668\") " pod="metallb-system/speaker-cc45k" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.173086 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqsc8\" (UniqueName: \"kubernetes.io/projected/9a498324-bbec-40e1-9649-f23ca26561f3-kube-api-access-bqsc8\") pod \"controller-5d688f5ffc-gnclz\" (UID: \"9a498324-bbec-40e1-9649-f23ca26561f3\") " pod="metallb-system/controller-5d688f5ffc-gnclz" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.173110 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/c6570c99-4cff-4194-9744-0efc00f88332-frr-conf\") pod \"frr-k8s-csxhg\" (UID: \"c6570c99-4cff-4194-9744-0efc00f88332\") " pod="metallb-system/frr-k8s-csxhg" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.173135 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668-memberlist\") pod \"speaker-cc45k\" (UID: \"c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668\") " pod="metallb-system/speaker-cc45k" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.173154 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9a498324-bbec-40e1-9649-f23ca26561f3-metrics-certs\") pod \"controller-5d688f5ffc-gnclz\" (UID: \"9a498324-bbec-40e1-9649-f23ca26561f3\") " pod="metallb-system/controller-5d688f5ffc-gnclz" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.173184 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9a498324-bbec-40e1-9649-f23ca26561f3-cert\") pod \"controller-5d688f5ffc-gnclz\" (UID: \"9a498324-bbec-40e1-9649-f23ca26561f3\") " pod="metallb-system/controller-5d688f5ffc-gnclz" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.173200 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/c6570c99-4cff-4194-9744-0efc00f88332-frr-startup\") pod \"frr-k8s-csxhg\" (UID: \"c6570c99-4cff-4194-9744-0efc00f88332\") " pod="metallb-system/frr-k8s-csxhg" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.173219 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dp8k\" (UniqueName: \"kubernetes.io/projected/c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668-kube-api-access-4dp8k\") pod \"speaker-cc45k\" (UID: \"c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668\") " pod="metallb-system/speaker-cc45k" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.173239 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/c6570c99-4cff-4194-9744-0efc00f88332-reloader\") pod \"frr-k8s-csxhg\" (UID: \"c6570c99-4cff-4194-9744-0efc00f88332\") " pod="metallb-system/frr-k8s-csxhg" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.173270 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/c6570c99-4cff-4194-9744-0efc00f88332-metrics\") pod \"frr-k8s-csxhg\" (UID: \"c6570c99-4cff-4194-9744-0efc00f88332\") " pod="metallb-system/frr-k8s-csxhg" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.173293 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668-metallb-excludel2\") pod \"speaker-cc45k\" (UID: \"c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668\") " pod="metallb-system/speaker-cc45k" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.173310 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/c6570c99-4cff-4194-9744-0efc00f88332-frr-sockets\") pod \"frr-k8s-csxhg\" (UID: \"c6570c99-4cff-4194-9744-0efc00f88332\") " pod="metallb-system/frr-k8s-csxhg" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.173328 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vj6p6\" (UniqueName: \"kubernetes.io/projected/c6570c99-4cff-4194-9744-0efc00f88332-kube-api-access-vj6p6\") pod \"frr-k8s-csxhg\" (UID: \"c6570c99-4cff-4194-9744-0efc00f88332\") " pod="metallb-system/frr-k8s-csxhg" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.174331 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/c6570c99-4cff-4194-9744-0efc00f88332-frr-conf\") pod \"frr-k8s-csxhg\" (UID: \"c6570c99-4cff-4194-9744-0efc00f88332\") " pod="metallb-system/frr-k8s-csxhg" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.174388 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/c6570c99-4cff-4194-9744-0efc00f88332-metrics\") pod \"frr-k8s-csxhg\" (UID: \"c6570c99-4cff-4194-9744-0efc00f88332\") " pod="metallb-system/frr-k8s-csxhg" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.174624 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/c6570c99-4cff-4194-9744-0efc00f88332-frr-sockets\") pod \"frr-k8s-csxhg\" (UID: \"c6570c99-4cff-4194-9744-0efc00f88332\") " pod="metallb-system/frr-k8s-csxhg" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.174980 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/c6570c99-4cff-4194-9744-0efc00f88332-frr-startup\") pod \"frr-k8s-csxhg\" (UID: \"c6570c99-4cff-4194-9744-0efc00f88332\") " pod="metallb-system/frr-k8s-csxhg" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.174986 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/c6570c99-4cff-4194-9744-0efc00f88332-reloader\") pod \"frr-k8s-csxhg\" (UID: \"c6570c99-4cff-4194-9744-0efc00f88332\") " pod="metallb-system/frr-k8s-csxhg" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.181199 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c6570c99-4cff-4194-9744-0efc00f88332-metrics-certs\") pod \"frr-k8s-csxhg\" (UID: \"c6570c99-4cff-4194-9744-0efc00f88332\") " pod="metallb-system/frr-k8s-csxhg" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.196915 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-kcxb8" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.210363 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vj6p6\" (UniqueName: \"kubernetes.io/projected/c6570c99-4cff-4194-9744-0efc00f88332-kube-api-access-vj6p6\") pod \"frr-k8s-csxhg\" (UID: \"c6570c99-4cff-4194-9744-0efc00f88332\") " pod="metallb-system/frr-k8s-csxhg" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.216470 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-csxhg" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.274927 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668-metrics-certs\") pod \"speaker-cc45k\" (UID: \"c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668\") " pod="metallb-system/speaker-cc45k" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.275623 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqsc8\" (UniqueName: \"kubernetes.io/projected/9a498324-bbec-40e1-9649-f23ca26561f3-kube-api-access-bqsc8\") pod \"controller-5d688f5ffc-gnclz\" (UID: \"9a498324-bbec-40e1-9649-f23ca26561f3\") " pod="metallb-system/controller-5d688f5ffc-gnclz" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.275672 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668-memberlist\") pod \"speaker-cc45k\" (UID: \"c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668\") " pod="metallb-system/speaker-cc45k" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.275702 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9a498324-bbec-40e1-9649-f23ca26561f3-metrics-certs\") pod \"controller-5d688f5ffc-gnclz\" (UID: \"9a498324-bbec-40e1-9649-f23ca26561f3\") " pod="metallb-system/controller-5d688f5ffc-gnclz" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.275742 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9a498324-bbec-40e1-9649-f23ca26561f3-cert\") pod \"controller-5d688f5ffc-gnclz\" (UID: \"9a498324-bbec-40e1-9649-f23ca26561f3\") " pod="metallb-system/controller-5d688f5ffc-gnclz" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.275765 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dp8k\" (UniqueName: \"kubernetes.io/projected/c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668-kube-api-access-4dp8k\") pod \"speaker-cc45k\" (UID: \"c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668\") " pod="metallb-system/speaker-cc45k" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.275809 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668-metallb-excludel2\") pod \"speaker-cc45k\" (UID: \"c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668\") " pod="metallb-system/speaker-cc45k" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.278248 4787 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.280347 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668-metrics-certs\") pod \"speaker-cc45k\" (UID: \"c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668\") " pod="metallb-system/speaker-cc45k" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.282236 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9a498324-bbec-40e1-9649-f23ca26561f3-metrics-certs\") pod \"controller-5d688f5ffc-gnclz\" (UID: \"9a498324-bbec-40e1-9649-f23ca26561f3\") " pod="metallb-system/controller-5d688f5ffc-gnclz" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.292065 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9a498324-bbec-40e1-9649-f23ca26561f3-cert\") pod \"controller-5d688f5ffc-gnclz\" (UID: \"9a498324-bbec-40e1-9649-f23ca26561f3\") " pod="metallb-system/controller-5d688f5ffc-gnclz" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.293411 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dp8k\" (UniqueName: \"kubernetes.io/projected/c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668-kube-api-access-4dp8k\") pod \"speaker-cc45k\" (UID: \"c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668\") " pod="metallb-system/speaker-cc45k" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.297090 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqsc8\" (UniqueName: \"kubernetes.io/projected/9a498324-bbec-40e1-9649-f23ca26561f3-kube-api-access-bqsc8\") pod \"controller-5d688f5ffc-gnclz\" (UID: \"9a498324-bbec-40e1-9649-f23ca26561f3\") " pod="metallb-system/controller-5d688f5ffc-gnclz" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.453670 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5d688f5ffc-gnclz" Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.619929 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-kcxb8"] Oct 01 09:48:19 crc kubenswrapper[4787]: W1001 09:48:19.633783 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8b191fd2_3066_4e07_9641_98e44be362f5.slice/crio-188a1bc7e9eb201638d65dd70d69c03f0cec1d376153941ad484469cc1aed997 WatchSource:0}: Error finding container 188a1bc7e9eb201638d65dd70d69c03f0cec1d376153941ad484469cc1aed997: Status 404 returned error can't find the container with id 188a1bc7e9eb201638d65dd70d69c03f0cec1d376153941ad484469cc1aed997 Oct 01 09:48:19 crc kubenswrapper[4787]: I1001 09:48:19.668993 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5d688f5ffc-gnclz"] Oct 01 09:48:19 crc kubenswrapper[4787]: W1001 09:48:19.673626 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9a498324_bbec_40e1_9649_f23ca26561f3.slice/crio-d256eff2dffe0be89acac006fafc62e9517235a47b08f26e6b0b9310eef8c713 WatchSource:0}: Error finding container d256eff2dffe0be89acac006fafc62e9517235a47b08f26e6b0b9310eef8c713: Status 404 returned error can't find the container with id d256eff2dffe0be89acac006fafc62e9517235a47b08f26e6b0b9310eef8c713 Oct 01 09:48:20 crc kubenswrapper[4787]: E1001 09:48:20.276482 4787 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: failed to sync secret cache: timed out waiting for the condition Oct 01 09:48:20 crc kubenswrapper[4787]: E1001 09:48:20.277974 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668-memberlist podName:c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668 nodeName:}" failed. No retries permitted until 2025-10-01 09:48:20.777953259 +0000 UTC m=+732.893097416 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668-memberlist") pod "speaker-cc45k" (UID: "c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668") : failed to sync secret cache: timed out waiting for the condition Oct 01 09:48:20 crc kubenswrapper[4787]: E1001 09:48:20.276621 4787 configmap.go:193] Couldn't get configMap metallb-system/metallb-excludel2: failed to sync configmap cache: timed out waiting for the condition Oct 01 09:48:20 crc kubenswrapper[4787]: E1001 09:48:20.278148 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668-metallb-excludel2 podName:c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668 nodeName:}" failed. No retries permitted until 2025-10-01 09:48:20.778140014 +0000 UTC m=+732.893284171 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metallb-excludel2" (UniqueName: "kubernetes.io/configmap/c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668-metallb-excludel2") pod "speaker-cc45k" (UID: "c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668") : failed to sync configmap cache: timed out waiting for the condition Oct 01 09:48:20 crc kubenswrapper[4787]: I1001 09:48:20.309413 4787 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 01 09:48:20 crc kubenswrapper[4787]: I1001 09:48:20.381901 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-csxhg" event={"ID":"c6570c99-4cff-4194-9744-0efc00f88332","Type":"ContainerStarted","Data":"159cebb26b11e5d089024761ab882894e69ea4da8a732cd2e71c2aaccc1eee83"} Oct 01 09:48:20 crc kubenswrapper[4787]: I1001 09:48:20.382550 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 01 09:48:20 crc kubenswrapper[4787]: I1001 09:48:20.384135 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5d688f5ffc-gnclz" event={"ID":"9a498324-bbec-40e1-9649-f23ca26561f3","Type":"ContainerStarted","Data":"ec6624c5634b88009cd0ea4aad10fcb58be7d1483889fca79080863025fd3c14"} Oct 01 09:48:20 crc kubenswrapper[4787]: I1001 09:48:20.384169 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5d688f5ffc-gnclz" event={"ID":"9a498324-bbec-40e1-9649-f23ca26561f3","Type":"ContainerStarted","Data":"8bf5eceaf271e25d157604971d00bea5c682091e89ad32c77eda6c8c0021c7df"} Oct 01 09:48:20 crc kubenswrapper[4787]: I1001 09:48:20.384184 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5d688f5ffc-gnclz" event={"ID":"9a498324-bbec-40e1-9649-f23ca26561f3","Type":"ContainerStarted","Data":"d256eff2dffe0be89acac006fafc62e9517235a47b08f26e6b0b9310eef8c713"} Oct 01 09:48:20 crc kubenswrapper[4787]: I1001 09:48:20.385155 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-5d688f5ffc-gnclz" Oct 01 09:48:20 crc kubenswrapper[4787]: I1001 09:48:20.386340 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-kcxb8" event={"ID":"8b191fd2-3066-4e07-9641-98e44be362f5","Type":"ContainerStarted","Data":"188a1bc7e9eb201638d65dd70d69c03f0cec1d376153941ad484469cc1aed997"} Oct 01 09:48:20 crc kubenswrapper[4787]: I1001 09:48:20.408847 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-5d688f5ffc-gnclz" podStartSLOduration=1.408825217 podStartE2EDuration="1.408825217s" podCreationTimestamp="2025-10-01 09:48:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:48:20.404014173 +0000 UTC m=+732.519158340" watchObservedRunningTime="2025-10-01 09:48:20.408825217 +0000 UTC m=+732.523969364" Oct 01 09:48:20 crc kubenswrapper[4787]: I1001 09:48:20.800713 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668-memberlist\") pod \"speaker-cc45k\" (UID: \"c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668\") " pod="metallb-system/speaker-cc45k" Oct 01 09:48:20 crc kubenswrapper[4787]: I1001 09:48:20.800855 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668-metallb-excludel2\") pod \"speaker-cc45k\" (UID: \"c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668\") " pod="metallb-system/speaker-cc45k" Oct 01 09:48:20 crc kubenswrapper[4787]: I1001 09:48:20.801601 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668-metallb-excludel2\") pod \"speaker-cc45k\" (UID: \"c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668\") " pod="metallb-system/speaker-cc45k" Oct 01 09:48:20 crc kubenswrapper[4787]: I1001 09:48:20.811825 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668-memberlist\") pod \"speaker-cc45k\" (UID: \"c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668\") " pod="metallb-system/speaker-cc45k" Oct 01 09:48:20 crc kubenswrapper[4787]: I1001 09:48:20.854622 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-cc45k" Oct 01 09:48:20 crc kubenswrapper[4787]: W1001 09:48:20.884629 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc9ad2dd2_b456_4e2c_a1d0_5c7b8664c668.slice/crio-d7a28aee273aaad42da8dc1300e9b6c4a96612dc8f847600d7d912221a9eb751 WatchSource:0}: Error finding container d7a28aee273aaad42da8dc1300e9b6c4a96612dc8f847600d7d912221a9eb751: Status 404 returned error can't find the container with id d7a28aee273aaad42da8dc1300e9b6c4a96612dc8f847600d7d912221a9eb751 Oct 01 09:48:21 crc kubenswrapper[4787]: I1001 09:48:21.399122 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-cc45k" event={"ID":"c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668","Type":"ContainerStarted","Data":"ed2a242463581885c0550912d80a4ad0e268ab6bf5e6cb0a5d39b79d892daf7b"} Oct 01 09:48:21 crc kubenswrapper[4787]: I1001 09:48:21.399533 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-cc45k" event={"ID":"c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668","Type":"ContainerStarted","Data":"d7a28aee273aaad42da8dc1300e9b6c4a96612dc8f847600d7d912221a9eb751"} Oct 01 09:48:22 crc kubenswrapper[4787]: I1001 09:48:22.430774 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-cc45k" event={"ID":"c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668","Type":"ContainerStarted","Data":"2ff0c1bd449c1e7d200902223f7f126775e032162f2b8340f383a7b15fb57780"} Oct 01 09:48:22 crc kubenswrapper[4787]: I1001 09:48:22.454812 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-cc45k" podStartSLOduration=3.45479741 podStartE2EDuration="3.45479741s" podCreationTimestamp="2025-10-01 09:48:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:48:22.453449885 +0000 UTC m=+734.568594042" watchObservedRunningTime="2025-10-01 09:48:22.45479741 +0000 UTC m=+734.569941567" Oct 01 09:48:23 crc kubenswrapper[4787]: I1001 09:48:23.437856 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-cc45k" Oct 01 09:48:27 crc kubenswrapper[4787]: I1001 09:48:27.481697 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-kcxb8" event={"ID":"8b191fd2-3066-4e07-9641-98e44be362f5","Type":"ContainerStarted","Data":"668077cf8c7336771d7032f0b64bfe6af4e4695577aab1b2a05da88ea365db45"} Oct 01 09:48:27 crc kubenswrapper[4787]: I1001 09:48:27.483253 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-kcxb8" Oct 01 09:48:27 crc kubenswrapper[4787]: I1001 09:48:27.483943 4787 generic.go:334] "Generic (PLEG): container finished" podID="c6570c99-4cff-4194-9744-0efc00f88332" containerID="6eb5503c357fcf839bc7b7cd2a834252a44815cc824b46565d1c4eb20e31a8f4" exitCode=0 Oct 01 09:48:27 crc kubenswrapper[4787]: I1001 09:48:27.483980 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-csxhg" event={"ID":"c6570c99-4cff-4194-9744-0efc00f88332","Type":"ContainerDied","Data":"6eb5503c357fcf839bc7b7cd2a834252a44815cc824b46565d1c4eb20e31a8f4"} Oct 01 09:48:27 crc kubenswrapper[4787]: I1001 09:48:27.500895 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-kcxb8" podStartSLOduration=2.365785639 podStartE2EDuration="9.500877017s" podCreationTimestamp="2025-10-01 09:48:18 +0000 UTC" firstStartedPulling="2025-10-01 09:48:19.636872554 +0000 UTC m=+731.752016711" lastFinishedPulling="2025-10-01 09:48:26.771963932 +0000 UTC m=+738.887108089" observedRunningTime="2025-10-01 09:48:27.496782772 +0000 UTC m=+739.611926969" watchObservedRunningTime="2025-10-01 09:48:27.500877017 +0000 UTC m=+739.616021194" Oct 01 09:48:28 crc kubenswrapper[4787]: I1001 09:48:28.493724 4787 generic.go:334] "Generic (PLEG): container finished" podID="c6570c99-4cff-4194-9744-0efc00f88332" containerID="4eaf9ec38847475cfac8debc851470d64be2c3e2fda0f29819701a8d40884b5c" exitCode=0 Oct 01 09:48:28 crc kubenswrapper[4787]: I1001 09:48:28.494660 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-csxhg" event={"ID":"c6570c99-4cff-4194-9744-0efc00f88332","Type":"ContainerDied","Data":"4eaf9ec38847475cfac8debc851470d64be2c3e2fda0f29819701a8d40884b5c"} Oct 01 09:48:29 crc kubenswrapper[4787]: I1001 09:48:29.457552 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-5d688f5ffc-gnclz" Oct 01 09:48:29 crc kubenswrapper[4787]: I1001 09:48:29.502367 4787 generic.go:334] "Generic (PLEG): container finished" podID="c6570c99-4cff-4194-9744-0efc00f88332" containerID="857d8e8283233ce5c95a3184f8fcfea01cfcb5f4ba5421ed5e5a76d00b3aa5a2" exitCode=0 Oct 01 09:48:29 crc kubenswrapper[4787]: I1001 09:48:29.502424 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-csxhg" event={"ID":"c6570c99-4cff-4194-9744-0efc00f88332","Type":"ContainerDied","Data":"857d8e8283233ce5c95a3184f8fcfea01cfcb5f4ba5421ed5e5a76d00b3aa5a2"} Oct 01 09:48:30 crc kubenswrapper[4787]: I1001 09:48:30.518778 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-csxhg" event={"ID":"c6570c99-4cff-4194-9744-0efc00f88332","Type":"ContainerStarted","Data":"6c104a9a3d27a87015ac8d0b826dfd4486fcf4cf763b34b577229e5ca8469e78"} Oct 01 09:48:30 crc kubenswrapper[4787]: I1001 09:48:30.519164 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-csxhg" event={"ID":"c6570c99-4cff-4194-9744-0efc00f88332","Type":"ContainerStarted","Data":"b15a25cfa658e29ba995bd7c3b3e8ad9c28b46b0606607a9aacd6bdf5544a23a"} Oct 01 09:48:30 crc kubenswrapper[4787]: I1001 09:48:30.519176 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-csxhg" event={"ID":"c6570c99-4cff-4194-9744-0efc00f88332","Type":"ContainerStarted","Data":"01defe80651d29d20ac2ebafccaa96f11aebddc79dd21f2b28c7564ae888804a"} Oct 01 09:48:30 crc kubenswrapper[4787]: I1001 09:48:30.519187 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-csxhg" event={"ID":"c6570c99-4cff-4194-9744-0efc00f88332","Type":"ContainerStarted","Data":"c74832797c108efea9e8beddd83c68ac59e9fa295cc35a9fc049b918e5682838"} Oct 01 09:48:30 crc kubenswrapper[4787]: I1001 09:48:30.519196 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-csxhg" event={"ID":"c6570c99-4cff-4194-9744-0efc00f88332","Type":"ContainerStarted","Data":"747b3409e99124a944a12d7f2446d4d5691ceee9c111958a4cc9f56d216da250"} Oct 01 09:48:31 crc kubenswrapper[4787]: I1001 09:48:31.532116 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-csxhg" event={"ID":"c6570c99-4cff-4194-9744-0efc00f88332","Type":"ContainerStarted","Data":"a900b4b1f4ddd440163ddb20b90ea643d4c78b23218779708339c4804ee769c8"} Oct 01 09:48:31 crc kubenswrapper[4787]: I1001 09:48:31.532614 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-csxhg" Oct 01 09:48:31 crc kubenswrapper[4787]: I1001 09:48:31.562414 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-csxhg" podStartSLOduration=6.163441645 podStartE2EDuration="13.562379561s" podCreationTimestamp="2025-10-01 09:48:18 +0000 UTC" firstStartedPulling="2025-10-01 09:48:19.375752446 +0000 UTC m=+731.490896603" lastFinishedPulling="2025-10-01 09:48:26.774690362 +0000 UTC m=+738.889834519" observedRunningTime="2025-10-01 09:48:31.55654403 +0000 UTC m=+743.671688267" watchObservedRunningTime="2025-10-01 09:48:31.562379561 +0000 UTC m=+743.677523758" Oct 01 09:48:33 crc kubenswrapper[4787]: I1001 09:48:33.391894 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-x649j"] Oct 01 09:48:33 crc kubenswrapper[4787]: I1001 09:48:33.392499 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-x649j" podUID="e6f1b5a2-581a-4c1a-b952-5662b2fb636f" containerName="controller-manager" containerID="cri-o://6d26013c5f8421594cca1529739132a4ce30cbc1f627f8e35431996d17074ba8" gracePeriod=30 Oct 01 09:48:33 crc kubenswrapper[4787]: I1001 09:48:33.540795 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-djtn2"] Oct 01 09:48:33 crc kubenswrapper[4787]: I1001 09:48:33.541115 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-djtn2" podUID="5f699a8f-ffe8-4512-9e55-24c88e4154d2" containerName="route-controller-manager" containerID="cri-o://89335eacca626ba7a51ed255e2110f680e83ab9eef027d90d08b322e5ee08fd8" gracePeriod=30 Oct 01 09:48:33 crc kubenswrapper[4787]: I1001 09:48:33.555476 4787 generic.go:334] "Generic (PLEG): container finished" podID="e6f1b5a2-581a-4c1a-b952-5662b2fb636f" containerID="6d26013c5f8421594cca1529739132a4ce30cbc1f627f8e35431996d17074ba8" exitCode=0 Oct 01 09:48:33 crc kubenswrapper[4787]: I1001 09:48:33.555564 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-x649j" event={"ID":"e6f1b5a2-581a-4c1a-b952-5662b2fb636f","Type":"ContainerDied","Data":"6d26013c5f8421594cca1529739132a4ce30cbc1f627f8e35431996d17074ba8"} Oct 01 09:48:33 crc kubenswrapper[4787]: I1001 09:48:33.837991 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-x649j" Oct 01 09:48:33 crc kubenswrapper[4787]: I1001 09:48:33.892884 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6f1b5a2-581a-4c1a-b952-5662b2fb636f-serving-cert\") pod \"e6f1b5a2-581a-4c1a-b952-5662b2fb636f\" (UID: \"e6f1b5a2-581a-4c1a-b952-5662b2fb636f\") " Oct 01 09:48:33 crc kubenswrapper[4787]: I1001 09:48:33.893011 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e6f1b5a2-581a-4c1a-b952-5662b2fb636f-client-ca\") pod \"e6f1b5a2-581a-4c1a-b952-5662b2fb636f\" (UID: \"e6f1b5a2-581a-4c1a-b952-5662b2fb636f\") " Oct 01 09:48:33 crc kubenswrapper[4787]: I1001 09:48:33.893058 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e6f1b5a2-581a-4c1a-b952-5662b2fb636f-proxy-ca-bundles\") pod \"e6f1b5a2-581a-4c1a-b952-5662b2fb636f\" (UID: \"e6f1b5a2-581a-4c1a-b952-5662b2fb636f\") " Oct 01 09:48:33 crc kubenswrapper[4787]: I1001 09:48:33.893194 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-628nj\" (UniqueName: \"kubernetes.io/projected/e6f1b5a2-581a-4c1a-b952-5662b2fb636f-kube-api-access-628nj\") pod \"e6f1b5a2-581a-4c1a-b952-5662b2fb636f\" (UID: \"e6f1b5a2-581a-4c1a-b952-5662b2fb636f\") " Oct 01 09:48:33 crc kubenswrapper[4787]: I1001 09:48:33.893353 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6f1b5a2-581a-4c1a-b952-5662b2fb636f-config\") pod \"e6f1b5a2-581a-4c1a-b952-5662b2fb636f\" (UID: \"e6f1b5a2-581a-4c1a-b952-5662b2fb636f\") " Oct 01 09:48:33 crc kubenswrapper[4787]: I1001 09:48:33.895014 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6f1b5a2-581a-4c1a-b952-5662b2fb636f-config" (OuterVolumeSpecName: "config") pod "e6f1b5a2-581a-4c1a-b952-5662b2fb636f" (UID: "e6f1b5a2-581a-4c1a-b952-5662b2fb636f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:48:33 crc kubenswrapper[4787]: I1001 09:48:33.895152 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6f1b5a2-581a-4c1a-b952-5662b2fb636f-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "e6f1b5a2-581a-4c1a-b952-5662b2fb636f" (UID: "e6f1b5a2-581a-4c1a-b952-5662b2fb636f"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:48:33 crc kubenswrapper[4787]: I1001 09:48:33.895438 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6f1b5a2-581a-4c1a-b952-5662b2fb636f-client-ca" (OuterVolumeSpecName: "client-ca") pod "e6f1b5a2-581a-4c1a-b952-5662b2fb636f" (UID: "e6f1b5a2-581a-4c1a-b952-5662b2fb636f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:48:33 crc kubenswrapper[4787]: I1001 09:48:33.900719 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6f1b5a2-581a-4c1a-b952-5662b2fb636f-kube-api-access-628nj" (OuterVolumeSpecName: "kube-api-access-628nj") pod "e6f1b5a2-581a-4c1a-b952-5662b2fb636f" (UID: "e6f1b5a2-581a-4c1a-b952-5662b2fb636f"). InnerVolumeSpecName "kube-api-access-628nj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:48:33 crc kubenswrapper[4787]: I1001 09:48:33.901326 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6f1b5a2-581a-4c1a-b952-5662b2fb636f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e6f1b5a2-581a-4c1a-b952-5662b2fb636f" (UID: "e6f1b5a2-581a-4c1a-b952-5662b2fb636f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:48:33 crc kubenswrapper[4787]: I1001 09:48:33.910807 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-djtn2" Oct 01 09:48:33 crc kubenswrapper[4787]: I1001 09:48:33.994499 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f699a8f-ffe8-4512-9e55-24c88e4154d2-config\") pod \"5f699a8f-ffe8-4512-9e55-24c88e4154d2\" (UID: \"5f699a8f-ffe8-4512-9e55-24c88e4154d2\") " Oct 01 09:48:33 crc kubenswrapper[4787]: I1001 09:48:33.994571 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89wcv\" (UniqueName: \"kubernetes.io/projected/5f699a8f-ffe8-4512-9e55-24c88e4154d2-kube-api-access-89wcv\") pod \"5f699a8f-ffe8-4512-9e55-24c88e4154d2\" (UID: \"5f699a8f-ffe8-4512-9e55-24c88e4154d2\") " Oct 01 09:48:33 crc kubenswrapper[4787]: I1001 09:48:33.994784 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5f699a8f-ffe8-4512-9e55-24c88e4154d2-client-ca\") pod \"5f699a8f-ffe8-4512-9e55-24c88e4154d2\" (UID: \"5f699a8f-ffe8-4512-9e55-24c88e4154d2\") " Oct 01 09:48:33 crc kubenswrapper[4787]: I1001 09:48:33.995697 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5f699a8f-ffe8-4512-9e55-24c88e4154d2-serving-cert\") pod \"5f699a8f-ffe8-4512-9e55-24c88e4154d2\" (UID: \"5f699a8f-ffe8-4512-9e55-24c88e4154d2\") " Oct 01 09:48:33 crc kubenswrapper[4787]: I1001 09:48:33.995809 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f699a8f-ffe8-4512-9e55-24c88e4154d2-client-ca" (OuterVolumeSpecName: "client-ca") pod "5f699a8f-ffe8-4512-9e55-24c88e4154d2" (UID: "5f699a8f-ffe8-4512-9e55-24c88e4154d2"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:48:33 crc kubenswrapper[4787]: I1001 09:48:33.996285 4787 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e6f1b5a2-581a-4c1a-b952-5662b2fb636f-client-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:48:33 crc kubenswrapper[4787]: I1001 09:48:33.996308 4787 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e6f1b5a2-581a-4c1a-b952-5662b2fb636f-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 01 09:48:33 crc kubenswrapper[4787]: I1001 09:48:33.996322 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-628nj\" (UniqueName: \"kubernetes.io/projected/e6f1b5a2-581a-4c1a-b952-5662b2fb636f-kube-api-access-628nj\") on node \"crc\" DevicePath \"\"" Oct 01 09:48:33 crc kubenswrapper[4787]: I1001 09:48:33.996333 4787 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5f699a8f-ffe8-4512-9e55-24c88e4154d2-client-ca\") on node \"crc\" DevicePath \"\"" Oct 01 09:48:33 crc kubenswrapper[4787]: I1001 09:48:33.996346 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6f1b5a2-581a-4c1a-b952-5662b2fb636f-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:48:33 crc kubenswrapper[4787]: I1001 09:48:33.996357 4787 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6f1b5a2-581a-4c1a-b952-5662b2fb636f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:48:33 crc kubenswrapper[4787]: I1001 09:48:33.996380 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f699a8f-ffe8-4512-9e55-24c88e4154d2-config" (OuterVolumeSpecName: "config") pod "5f699a8f-ffe8-4512-9e55-24c88e4154d2" (UID: "5f699a8f-ffe8-4512-9e55-24c88e4154d2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:48:33 crc kubenswrapper[4787]: I1001 09:48:33.999227 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f699a8f-ffe8-4512-9e55-24c88e4154d2-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5f699a8f-ffe8-4512-9e55-24c88e4154d2" (UID: "5f699a8f-ffe8-4512-9e55-24c88e4154d2"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:48:33 crc kubenswrapper[4787]: I1001 09:48:33.999280 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f699a8f-ffe8-4512-9e55-24c88e4154d2-kube-api-access-89wcv" (OuterVolumeSpecName: "kube-api-access-89wcv") pod "5f699a8f-ffe8-4512-9e55-24c88e4154d2" (UID: "5f699a8f-ffe8-4512-9e55-24c88e4154d2"). InnerVolumeSpecName "kube-api-access-89wcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:48:34 crc kubenswrapper[4787]: I1001 09:48:34.098362 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f699a8f-ffe8-4512-9e55-24c88e4154d2-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:48:34 crc kubenswrapper[4787]: I1001 09:48:34.098884 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89wcv\" (UniqueName: \"kubernetes.io/projected/5f699a8f-ffe8-4512-9e55-24c88e4154d2-kube-api-access-89wcv\") on node \"crc\" DevicePath \"\"" Oct 01 09:48:34 crc kubenswrapper[4787]: I1001 09:48:34.098909 4787 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5f699a8f-ffe8-4512-9e55-24c88e4154d2-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 01 09:48:34 crc kubenswrapper[4787]: I1001 09:48:34.217797 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-csxhg" Oct 01 09:48:34 crc kubenswrapper[4787]: I1001 09:48:34.255684 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-csxhg" Oct 01 09:48:34 crc kubenswrapper[4787]: I1001 09:48:34.563311 4787 generic.go:334] "Generic (PLEG): container finished" podID="5f699a8f-ffe8-4512-9e55-24c88e4154d2" containerID="89335eacca626ba7a51ed255e2110f680e83ab9eef027d90d08b322e5ee08fd8" exitCode=0 Oct 01 09:48:34 crc kubenswrapper[4787]: I1001 09:48:34.563376 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-djtn2" Oct 01 09:48:34 crc kubenswrapper[4787]: I1001 09:48:34.563383 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-djtn2" event={"ID":"5f699a8f-ffe8-4512-9e55-24c88e4154d2","Type":"ContainerDied","Data":"89335eacca626ba7a51ed255e2110f680e83ab9eef027d90d08b322e5ee08fd8"} Oct 01 09:48:34 crc kubenswrapper[4787]: I1001 09:48:34.563487 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-djtn2" event={"ID":"5f699a8f-ffe8-4512-9e55-24c88e4154d2","Type":"ContainerDied","Data":"63ef6ecd235c2be02d075ee4eb6afdff48eeb3978126c398915cca97cfb38c7a"} Oct 01 09:48:34 crc kubenswrapper[4787]: I1001 09:48:34.563507 4787 scope.go:117] "RemoveContainer" containerID="89335eacca626ba7a51ed255e2110f680e83ab9eef027d90d08b322e5ee08fd8" Oct 01 09:48:34 crc kubenswrapper[4787]: I1001 09:48:34.570711 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-x649j" event={"ID":"e6f1b5a2-581a-4c1a-b952-5662b2fb636f","Type":"ContainerDied","Data":"476b72cc982d783e926c5708376f78357f203fd47d3076351abd671992867c3d"} Oct 01 09:48:34 crc kubenswrapper[4787]: I1001 09:48:34.570722 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-x649j" Oct 01 09:48:34 crc kubenswrapper[4787]: I1001 09:48:34.587096 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-djtn2"] Oct 01 09:48:34 crc kubenswrapper[4787]: I1001 09:48:34.591448 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-djtn2"] Oct 01 09:48:34 crc kubenswrapper[4787]: I1001 09:48:34.603611 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-x649j"] Oct 01 09:48:34 crc kubenswrapper[4787]: I1001 09:48:34.609614 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-x649j"] Oct 01 09:48:34 crc kubenswrapper[4787]: I1001 09:48:34.898984 4787 scope.go:117] "RemoveContainer" containerID="89335eacca626ba7a51ed255e2110f680e83ab9eef027d90d08b322e5ee08fd8" Oct 01 09:48:34 crc kubenswrapper[4787]: E1001 09:48:34.899420 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89335eacca626ba7a51ed255e2110f680e83ab9eef027d90d08b322e5ee08fd8\": container with ID starting with 89335eacca626ba7a51ed255e2110f680e83ab9eef027d90d08b322e5ee08fd8 not found: ID does not exist" containerID="89335eacca626ba7a51ed255e2110f680e83ab9eef027d90d08b322e5ee08fd8" Oct 01 09:48:34 crc kubenswrapper[4787]: I1001 09:48:34.899466 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89335eacca626ba7a51ed255e2110f680e83ab9eef027d90d08b322e5ee08fd8"} err="failed to get container status \"89335eacca626ba7a51ed255e2110f680e83ab9eef027d90d08b322e5ee08fd8\": rpc error: code = NotFound desc = could not find container \"89335eacca626ba7a51ed255e2110f680e83ab9eef027d90d08b322e5ee08fd8\": container with ID starting with 89335eacca626ba7a51ed255e2110f680e83ab9eef027d90d08b322e5ee08fd8 not found: ID does not exist" Oct 01 09:48:34 crc kubenswrapper[4787]: I1001 09:48:34.899494 4787 scope.go:117] "RemoveContainer" containerID="6d26013c5f8421594cca1529739132a4ce30cbc1f627f8e35431996d17074ba8" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.467341 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-669b8578c-smxv9"] Oct 01 09:48:35 crc kubenswrapper[4787]: E1001 09:48:35.467627 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f699a8f-ffe8-4512-9e55-24c88e4154d2" containerName="route-controller-manager" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.467644 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f699a8f-ffe8-4512-9e55-24c88e4154d2" containerName="route-controller-manager" Oct 01 09:48:35 crc kubenswrapper[4787]: E1001 09:48:35.467676 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6f1b5a2-581a-4c1a-b952-5662b2fb636f" containerName="controller-manager" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.467685 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6f1b5a2-581a-4c1a-b952-5662b2fb636f" containerName="controller-manager" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.467841 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6f1b5a2-581a-4c1a-b952-5662b2fb636f" containerName="controller-manager" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.467860 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f699a8f-ffe8-4512-9e55-24c88e4154d2" containerName="route-controller-manager" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.468293 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-669b8578c-smxv9" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.470302 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-b58696bcf-zzqpd"] Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.470986 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-b58696bcf-zzqpd" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.471123 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.471387 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.472525 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.472872 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.472936 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.473678 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.473687 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.473831 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.474380 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.474384 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.474413 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.474839 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.494653 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.499799 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-669b8578c-smxv9"] Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.511345 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-b58696bcf-zzqpd"] Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.618704 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stvrl\" (UniqueName: \"kubernetes.io/projected/248a5afe-05e7-4942-9877-b7ac8d39faed-kube-api-access-stvrl\") pod \"route-controller-manager-b58696bcf-zzqpd\" (UID: \"248a5afe-05e7-4942-9877-b7ac8d39faed\") " pod="openshift-route-controller-manager/route-controller-manager-b58696bcf-zzqpd" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.618776 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3ef33ff-2dc0-4911-9af6-fa6d8909dc6f-config\") pod \"controller-manager-669b8578c-smxv9\" (UID: \"c3ef33ff-2dc0-4911-9af6-fa6d8909dc6f\") " pod="openshift-controller-manager/controller-manager-669b8578c-smxv9" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.618811 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/248a5afe-05e7-4942-9877-b7ac8d39faed-serving-cert\") pod \"route-controller-manager-b58696bcf-zzqpd\" (UID: \"248a5afe-05e7-4942-9877-b7ac8d39faed\") " pod="openshift-route-controller-manager/route-controller-manager-b58696bcf-zzqpd" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.618957 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bx2fk\" (UniqueName: \"kubernetes.io/projected/c3ef33ff-2dc0-4911-9af6-fa6d8909dc6f-kube-api-access-bx2fk\") pod \"controller-manager-669b8578c-smxv9\" (UID: \"c3ef33ff-2dc0-4911-9af6-fa6d8909dc6f\") " pod="openshift-controller-manager/controller-manager-669b8578c-smxv9" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.619035 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c3ef33ff-2dc0-4911-9af6-fa6d8909dc6f-client-ca\") pod \"controller-manager-669b8578c-smxv9\" (UID: \"c3ef33ff-2dc0-4911-9af6-fa6d8909dc6f\") " pod="openshift-controller-manager/controller-manager-669b8578c-smxv9" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.619205 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c3ef33ff-2dc0-4911-9af6-fa6d8909dc6f-serving-cert\") pod \"controller-manager-669b8578c-smxv9\" (UID: \"c3ef33ff-2dc0-4911-9af6-fa6d8909dc6f\") " pod="openshift-controller-manager/controller-manager-669b8578c-smxv9" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.619328 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/248a5afe-05e7-4942-9877-b7ac8d39faed-client-ca\") pod \"route-controller-manager-b58696bcf-zzqpd\" (UID: \"248a5afe-05e7-4942-9877-b7ac8d39faed\") " pod="openshift-route-controller-manager/route-controller-manager-b58696bcf-zzqpd" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.619471 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c3ef33ff-2dc0-4911-9af6-fa6d8909dc6f-proxy-ca-bundles\") pod \"controller-manager-669b8578c-smxv9\" (UID: \"c3ef33ff-2dc0-4911-9af6-fa6d8909dc6f\") " pod="openshift-controller-manager/controller-manager-669b8578c-smxv9" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.619495 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/248a5afe-05e7-4942-9877-b7ac8d39faed-config\") pod \"route-controller-manager-b58696bcf-zzqpd\" (UID: \"248a5afe-05e7-4942-9877-b7ac8d39faed\") " pod="openshift-route-controller-manager/route-controller-manager-b58696bcf-zzqpd" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.721704 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c3ef33ff-2dc0-4911-9af6-fa6d8909dc6f-proxy-ca-bundles\") pod \"controller-manager-669b8578c-smxv9\" (UID: \"c3ef33ff-2dc0-4911-9af6-fa6d8909dc6f\") " pod="openshift-controller-manager/controller-manager-669b8578c-smxv9" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.721798 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/248a5afe-05e7-4942-9877-b7ac8d39faed-config\") pod \"route-controller-manager-b58696bcf-zzqpd\" (UID: \"248a5afe-05e7-4942-9877-b7ac8d39faed\") " pod="openshift-route-controller-manager/route-controller-manager-b58696bcf-zzqpd" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.721840 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stvrl\" (UniqueName: \"kubernetes.io/projected/248a5afe-05e7-4942-9877-b7ac8d39faed-kube-api-access-stvrl\") pod \"route-controller-manager-b58696bcf-zzqpd\" (UID: \"248a5afe-05e7-4942-9877-b7ac8d39faed\") " pod="openshift-route-controller-manager/route-controller-manager-b58696bcf-zzqpd" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.721880 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3ef33ff-2dc0-4911-9af6-fa6d8909dc6f-config\") pod \"controller-manager-669b8578c-smxv9\" (UID: \"c3ef33ff-2dc0-4911-9af6-fa6d8909dc6f\") " pod="openshift-controller-manager/controller-manager-669b8578c-smxv9" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.721933 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/248a5afe-05e7-4942-9877-b7ac8d39faed-serving-cert\") pod \"route-controller-manager-b58696bcf-zzqpd\" (UID: \"248a5afe-05e7-4942-9877-b7ac8d39faed\") " pod="openshift-route-controller-manager/route-controller-manager-b58696bcf-zzqpd" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.722024 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bx2fk\" (UniqueName: \"kubernetes.io/projected/c3ef33ff-2dc0-4911-9af6-fa6d8909dc6f-kube-api-access-bx2fk\") pod \"controller-manager-669b8578c-smxv9\" (UID: \"c3ef33ff-2dc0-4911-9af6-fa6d8909dc6f\") " pod="openshift-controller-manager/controller-manager-669b8578c-smxv9" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.722120 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c3ef33ff-2dc0-4911-9af6-fa6d8909dc6f-client-ca\") pod \"controller-manager-669b8578c-smxv9\" (UID: \"c3ef33ff-2dc0-4911-9af6-fa6d8909dc6f\") " pod="openshift-controller-manager/controller-manager-669b8578c-smxv9" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.722210 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c3ef33ff-2dc0-4911-9af6-fa6d8909dc6f-serving-cert\") pod \"controller-manager-669b8578c-smxv9\" (UID: \"c3ef33ff-2dc0-4911-9af6-fa6d8909dc6f\") " pod="openshift-controller-manager/controller-manager-669b8578c-smxv9" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.722312 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/248a5afe-05e7-4942-9877-b7ac8d39faed-client-ca\") pod \"route-controller-manager-b58696bcf-zzqpd\" (UID: \"248a5afe-05e7-4942-9877-b7ac8d39faed\") " pod="openshift-route-controller-manager/route-controller-manager-b58696bcf-zzqpd" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.723192 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c3ef33ff-2dc0-4911-9af6-fa6d8909dc6f-proxy-ca-bundles\") pod \"controller-manager-669b8578c-smxv9\" (UID: \"c3ef33ff-2dc0-4911-9af6-fa6d8909dc6f\") " pod="openshift-controller-manager/controller-manager-669b8578c-smxv9" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.723235 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c3ef33ff-2dc0-4911-9af6-fa6d8909dc6f-client-ca\") pod \"controller-manager-669b8578c-smxv9\" (UID: \"c3ef33ff-2dc0-4911-9af6-fa6d8909dc6f\") " pod="openshift-controller-manager/controller-manager-669b8578c-smxv9" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.723553 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/248a5afe-05e7-4942-9877-b7ac8d39faed-config\") pod \"route-controller-manager-b58696bcf-zzqpd\" (UID: \"248a5afe-05e7-4942-9877-b7ac8d39faed\") " pod="openshift-route-controller-manager/route-controller-manager-b58696bcf-zzqpd" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.723581 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3ef33ff-2dc0-4911-9af6-fa6d8909dc6f-config\") pod \"controller-manager-669b8578c-smxv9\" (UID: \"c3ef33ff-2dc0-4911-9af6-fa6d8909dc6f\") " pod="openshift-controller-manager/controller-manager-669b8578c-smxv9" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.724576 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/248a5afe-05e7-4942-9877-b7ac8d39faed-client-ca\") pod \"route-controller-manager-b58696bcf-zzqpd\" (UID: \"248a5afe-05e7-4942-9877-b7ac8d39faed\") " pod="openshift-route-controller-manager/route-controller-manager-b58696bcf-zzqpd" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.728031 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/248a5afe-05e7-4942-9877-b7ac8d39faed-serving-cert\") pod \"route-controller-manager-b58696bcf-zzqpd\" (UID: \"248a5afe-05e7-4942-9877-b7ac8d39faed\") " pod="openshift-route-controller-manager/route-controller-manager-b58696bcf-zzqpd" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.729065 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c3ef33ff-2dc0-4911-9af6-fa6d8909dc6f-serving-cert\") pod \"controller-manager-669b8578c-smxv9\" (UID: \"c3ef33ff-2dc0-4911-9af6-fa6d8909dc6f\") " pod="openshift-controller-manager/controller-manager-669b8578c-smxv9" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.742278 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bx2fk\" (UniqueName: \"kubernetes.io/projected/c3ef33ff-2dc0-4911-9af6-fa6d8909dc6f-kube-api-access-bx2fk\") pod \"controller-manager-669b8578c-smxv9\" (UID: \"c3ef33ff-2dc0-4911-9af6-fa6d8909dc6f\") " pod="openshift-controller-manager/controller-manager-669b8578c-smxv9" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.749434 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stvrl\" (UniqueName: \"kubernetes.io/projected/248a5afe-05e7-4942-9877-b7ac8d39faed-kube-api-access-stvrl\") pod \"route-controller-manager-b58696bcf-zzqpd\" (UID: \"248a5afe-05e7-4942-9877-b7ac8d39faed\") " pod="openshift-route-controller-manager/route-controller-manager-b58696bcf-zzqpd" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.790546 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-669b8578c-smxv9" Oct 01 09:48:35 crc kubenswrapper[4787]: I1001 09:48:35.803163 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-b58696bcf-zzqpd" Oct 01 09:48:36 crc kubenswrapper[4787]: I1001 09:48:36.029822 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-669b8578c-smxv9"] Oct 01 09:48:36 crc kubenswrapper[4787]: W1001 09:48:36.038869 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc3ef33ff_2dc0_4911_9af6_fa6d8909dc6f.slice/crio-ef76e1bf284632b848f824549af2cd301d2956693f56d7416e5160ebdf92f8dc WatchSource:0}: Error finding container ef76e1bf284632b848f824549af2cd301d2956693f56d7416e5160ebdf92f8dc: Status 404 returned error can't find the container with id ef76e1bf284632b848f824549af2cd301d2956693f56d7416e5160ebdf92f8dc Oct 01 09:48:36 crc kubenswrapper[4787]: I1001 09:48:36.268198 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-b58696bcf-zzqpd"] Oct 01 09:48:36 crc kubenswrapper[4787]: W1001 09:48:36.273902 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod248a5afe_05e7_4942_9877_b7ac8d39faed.slice/crio-402adf0e795c29b0db87880b83d26dc40fe2c38d5919577f0da6547cf87f54f6 WatchSource:0}: Error finding container 402adf0e795c29b0db87880b83d26dc40fe2c38d5919577f0da6547cf87f54f6: Status 404 returned error can't find the container with id 402adf0e795c29b0db87880b83d26dc40fe2c38d5919577f0da6547cf87f54f6 Oct 01 09:48:36 crc kubenswrapper[4787]: I1001 09:48:36.532310 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f699a8f-ffe8-4512-9e55-24c88e4154d2" path="/var/lib/kubelet/pods/5f699a8f-ffe8-4512-9e55-24c88e4154d2/volumes" Oct 01 09:48:36 crc kubenswrapper[4787]: I1001 09:48:36.533043 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6f1b5a2-581a-4c1a-b952-5662b2fb636f" path="/var/lib/kubelet/pods/e6f1b5a2-581a-4c1a-b952-5662b2fb636f/volumes" Oct 01 09:48:36 crc kubenswrapper[4787]: I1001 09:48:36.586963 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-b58696bcf-zzqpd" event={"ID":"248a5afe-05e7-4942-9877-b7ac8d39faed","Type":"ContainerStarted","Data":"f87cb948a80cd16635cc3752de120d08e63ee5d950161c5c8f292e8e0092bd1b"} Oct 01 09:48:36 crc kubenswrapper[4787]: I1001 09:48:36.587096 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-b58696bcf-zzqpd" event={"ID":"248a5afe-05e7-4942-9877-b7ac8d39faed","Type":"ContainerStarted","Data":"402adf0e795c29b0db87880b83d26dc40fe2c38d5919577f0da6547cf87f54f6"} Oct 01 09:48:36 crc kubenswrapper[4787]: I1001 09:48:36.587120 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-b58696bcf-zzqpd" Oct 01 09:48:36 crc kubenswrapper[4787]: I1001 09:48:36.589727 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-669b8578c-smxv9" event={"ID":"c3ef33ff-2dc0-4911-9af6-fa6d8909dc6f","Type":"ContainerStarted","Data":"65602d9327caa278136b0796d7b872dc9377f6229e16bb66a72cad09b89d6143"} Oct 01 09:48:36 crc kubenswrapper[4787]: I1001 09:48:36.589763 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-669b8578c-smxv9" event={"ID":"c3ef33ff-2dc0-4911-9af6-fa6d8909dc6f","Type":"ContainerStarted","Data":"ef76e1bf284632b848f824549af2cd301d2956693f56d7416e5160ebdf92f8dc"} Oct 01 09:48:36 crc kubenswrapper[4787]: I1001 09:48:36.590046 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-669b8578c-smxv9" Oct 01 09:48:36 crc kubenswrapper[4787]: I1001 09:48:36.595706 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-669b8578c-smxv9" Oct 01 09:48:36 crc kubenswrapper[4787]: I1001 09:48:36.613096 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-b58696bcf-zzqpd" podStartSLOduration=3.613051746 podStartE2EDuration="3.613051746s" podCreationTimestamp="2025-10-01 09:48:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:48:36.607698188 +0000 UTC m=+748.722842365" watchObservedRunningTime="2025-10-01 09:48:36.613051746 +0000 UTC m=+748.728195903" Oct 01 09:48:36 crc kubenswrapper[4787]: I1001 09:48:36.632980 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-669b8578c-smxv9" podStartSLOduration=3.632960988 podStartE2EDuration="3.632960988s" podCreationTimestamp="2025-10-01 09:48:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:48:36.630251658 +0000 UTC m=+748.745395815" watchObservedRunningTime="2025-10-01 09:48:36.632960988 +0000 UTC m=+748.748105145" Oct 01 09:48:36 crc kubenswrapper[4787]: I1001 09:48:36.785649 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-b58696bcf-zzqpd" Oct 01 09:48:39 crc kubenswrapper[4787]: I1001 09:48:39.203411 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-kcxb8" Oct 01 09:48:39 crc kubenswrapper[4787]: I1001 09:48:39.222464 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-csxhg" Oct 01 09:48:40 crc kubenswrapper[4787]: I1001 09:48:40.859580 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-cc45k" Oct 01 09:48:41 crc kubenswrapper[4787]: I1001 09:48:41.250428 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:48:41 crc kubenswrapper[4787]: I1001 09:48:41.250497 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:48:41 crc kubenswrapper[4787]: I1001 09:48:41.250545 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" Oct 01 09:48:41 crc kubenswrapper[4787]: I1001 09:48:41.251158 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c5956c6b3dfb9dae0b884bab2812d62e85ed0b6d8154a894d187e5889824b51f"} pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 09:48:41 crc kubenswrapper[4787]: I1001 09:48:41.251229 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" containerID="cri-o://c5956c6b3dfb9dae0b884bab2812d62e85ed0b6d8154a894d187e5889824b51f" gracePeriod=600 Oct 01 09:48:41 crc kubenswrapper[4787]: I1001 09:48:41.619934 4787 generic.go:334] "Generic (PLEG): container finished" podID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerID="c5956c6b3dfb9dae0b884bab2812d62e85ed0b6d8154a894d187e5889824b51f" exitCode=0 Oct 01 09:48:41 crc kubenswrapper[4787]: I1001 09:48:41.620104 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" event={"ID":"fa51b883-5f3e-4141-9d4c-37704aac7718","Type":"ContainerDied","Data":"c5956c6b3dfb9dae0b884bab2812d62e85ed0b6d8154a894d187e5889824b51f"} Oct 01 09:48:41 crc kubenswrapper[4787]: I1001 09:48:41.620915 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" event={"ID":"fa51b883-5f3e-4141-9d4c-37704aac7718","Type":"ContainerStarted","Data":"d423136db78a796507e81b181b34f0022442acd501d56cba0addad4da7455409"} Oct 01 09:48:41 crc kubenswrapper[4787]: I1001 09:48:41.621060 4787 scope.go:117] "RemoveContainer" containerID="e4d458228b35af45ba3f97cae60443c240914647056563a211cb5a9e256ce4f6" Oct 01 09:48:43 crc kubenswrapper[4787]: I1001 09:48:43.839990 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-wmqqd"] Oct 01 09:48:43 crc kubenswrapper[4787]: I1001 09:48:43.842235 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-wmqqd" Oct 01 09:48:43 crc kubenswrapper[4787]: I1001 09:48:43.851317 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 01 09:48:43 crc kubenswrapper[4787]: I1001 09:48:43.852701 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 01 09:48:43 crc kubenswrapper[4787]: I1001 09:48:43.859190 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-wmqqd"] Oct 01 09:48:43 crc kubenswrapper[4787]: I1001 09:48:43.945190 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnhn6\" (UniqueName: \"kubernetes.io/projected/be674cf7-d26e-464e-b661-4bff851d0413-kube-api-access-gnhn6\") pod \"openstack-operator-index-wmqqd\" (UID: \"be674cf7-d26e-464e-b661-4bff851d0413\") " pod="openstack-operators/openstack-operator-index-wmqqd" Oct 01 09:48:43 crc kubenswrapper[4787]: I1001 09:48:43.953227 4787 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 01 09:48:44 crc kubenswrapper[4787]: I1001 09:48:44.046536 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnhn6\" (UniqueName: \"kubernetes.io/projected/be674cf7-d26e-464e-b661-4bff851d0413-kube-api-access-gnhn6\") pod \"openstack-operator-index-wmqqd\" (UID: \"be674cf7-d26e-464e-b661-4bff851d0413\") " pod="openstack-operators/openstack-operator-index-wmqqd" Oct 01 09:48:44 crc kubenswrapper[4787]: I1001 09:48:44.071674 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnhn6\" (UniqueName: \"kubernetes.io/projected/be674cf7-d26e-464e-b661-4bff851d0413-kube-api-access-gnhn6\") pod \"openstack-operator-index-wmqqd\" (UID: \"be674cf7-d26e-464e-b661-4bff851d0413\") " pod="openstack-operators/openstack-operator-index-wmqqd" Oct 01 09:48:44 crc kubenswrapper[4787]: I1001 09:48:44.173746 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-wmqqd" Oct 01 09:48:44 crc kubenswrapper[4787]: I1001 09:48:44.584909 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-wmqqd"] Oct 01 09:48:44 crc kubenswrapper[4787]: W1001 09:48:44.591618 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbe674cf7_d26e_464e_b661_4bff851d0413.slice/crio-0b72bec6e8b01ba95460ebfa652b73c4cd6c094f961ac7cf69f9a6b0e059471a WatchSource:0}: Error finding container 0b72bec6e8b01ba95460ebfa652b73c4cd6c094f961ac7cf69f9a6b0e059471a: Status 404 returned error can't find the container with id 0b72bec6e8b01ba95460ebfa652b73c4cd6c094f961ac7cf69f9a6b0e059471a Oct 01 09:48:44 crc kubenswrapper[4787]: I1001 09:48:44.650643 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-wmqqd" event={"ID":"be674cf7-d26e-464e-b661-4bff851d0413","Type":"ContainerStarted","Data":"0b72bec6e8b01ba95460ebfa652b73c4cd6c094f961ac7cf69f9a6b0e059471a"} Oct 01 09:48:45 crc kubenswrapper[4787]: I1001 09:48:45.660409 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-wmqqd" event={"ID":"be674cf7-d26e-464e-b661-4bff851d0413","Type":"ContainerStarted","Data":"f84caac49271c19bcf1d172d60b7461e37f25120ee1add03510e9ee788c78225"} Oct 01 09:48:45 crc kubenswrapper[4787]: I1001 09:48:45.682208 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-wmqqd" podStartSLOduration=1.923799973 podStartE2EDuration="2.682171856s" podCreationTimestamp="2025-10-01 09:48:43 +0000 UTC" firstStartedPulling="2025-10-01 09:48:44.594352276 +0000 UTC m=+756.709496433" lastFinishedPulling="2025-10-01 09:48:45.352724159 +0000 UTC m=+757.467868316" observedRunningTime="2025-10-01 09:48:45.678696497 +0000 UTC m=+757.793840694" watchObservedRunningTime="2025-10-01 09:48:45.682171856 +0000 UTC m=+757.797316053" Oct 01 09:48:46 crc kubenswrapper[4787]: I1001 09:48:46.413703 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-wmqqd"] Oct 01 09:48:47 crc kubenswrapper[4787]: I1001 09:48:47.026177 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-jnbw4"] Oct 01 09:48:47 crc kubenswrapper[4787]: I1001 09:48:47.028368 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-jnbw4" Oct 01 09:48:47 crc kubenswrapper[4787]: I1001 09:48:47.046303 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-b2cnp" Oct 01 09:48:47 crc kubenswrapper[4787]: I1001 09:48:47.058888 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-jnbw4"] Oct 01 09:48:47 crc kubenswrapper[4787]: I1001 09:48:47.092397 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnm9r\" (UniqueName: \"kubernetes.io/projected/c6b4eeb7-99c2-44ff-9f71-e27c17c17649-kube-api-access-qnm9r\") pod \"openstack-operator-index-jnbw4\" (UID: \"c6b4eeb7-99c2-44ff-9f71-e27c17c17649\") " pod="openstack-operators/openstack-operator-index-jnbw4" Oct 01 09:48:47 crc kubenswrapper[4787]: I1001 09:48:47.194873 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnm9r\" (UniqueName: \"kubernetes.io/projected/c6b4eeb7-99c2-44ff-9f71-e27c17c17649-kube-api-access-qnm9r\") pod \"openstack-operator-index-jnbw4\" (UID: \"c6b4eeb7-99c2-44ff-9f71-e27c17c17649\") " pod="openstack-operators/openstack-operator-index-jnbw4" Oct 01 09:48:47 crc kubenswrapper[4787]: I1001 09:48:47.218533 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnm9r\" (UniqueName: \"kubernetes.io/projected/c6b4eeb7-99c2-44ff-9f71-e27c17c17649-kube-api-access-qnm9r\") pod \"openstack-operator-index-jnbw4\" (UID: \"c6b4eeb7-99c2-44ff-9f71-e27c17c17649\") " pod="openstack-operators/openstack-operator-index-jnbw4" Oct 01 09:48:47 crc kubenswrapper[4787]: I1001 09:48:47.357549 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-jnbw4" Oct 01 09:48:47 crc kubenswrapper[4787]: I1001 09:48:47.676312 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-wmqqd" podUID="be674cf7-d26e-464e-b661-4bff851d0413" containerName="registry-server" containerID="cri-o://f84caac49271c19bcf1d172d60b7461e37f25120ee1add03510e9ee788c78225" gracePeriod=2 Oct 01 09:48:47 crc kubenswrapper[4787]: I1001 09:48:47.807357 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-jnbw4"] Oct 01 09:48:47 crc kubenswrapper[4787]: W1001 09:48:47.818374 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc6b4eeb7_99c2_44ff_9f71_e27c17c17649.slice/crio-f5abf28b9b2de221d2902b1e80bf30b84d3921fe47492440efb2bb69a7372a5d WatchSource:0}: Error finding container f5abf28b9b2de221d2902b1e80bf30b84d3921fe47492440efb2bb69a7372a5d: Status 404 returned error can't find the container with id f5abf28b9b2de221d2902b1e80bf30b84d3921fe47492440efb2bb69a7372a5d Oct 01 09:48:48 crc kubenswrapper[4787]: I1001 09:48:48.065396 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-wmqqd" Oct 01 09:48:48 crc kubenswrapper[4787]: I1001 09:48:48.208973 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gnhn6\" (UniqueName: \"kubernetes.io/projected/be674cf7-d26e-464e-b661-4bff851d0413-kube-api-access-gnhn6\") pod \"be674cf7-d26e-464e-b661-4bff851d0413\" (UID: \"be674cf7-d26e-464e-b661-4bff851d0413\") " Oct 01 09:48:48 crc kubenswrapper[4787]: I1001 09:48:48.216946 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be674cf7-d26e-464e-b661-4bff851d0413-kube-api-access-gnhn6" (OuterVolumeSpecName: "kube-api-access-gnhn6") pod "be674cf7-d26e-464e-b661-4bff851d0413" (UID: "be674cf7-d26e-464e-b661-4bff851d0413"). InnerVolumeSpecName "kube-api-access-gnhn6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:48:48 crc kubenswrapper[4787]: I1001 09:48:48.310265 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gnhn6\" (UniqueName: \"kubernetes.io/projected/be674cf7-d26e-464e-b661-4bff851d0413-kube-api-access-gnhn6\") on node \"crc\" DevicePath \"\"" Oct 01 09:48:48 crc kubenswrapper[4787]: I1001 09:48:48.685017 4787 generic.go:334] "Generic (PLEG): container finished" podID="be674cf7-d26e-464e-b661-4bff851d0413" containerID="f84caac49271c19bcf1d172d60b7461e37f25120ee1add03510e9ee788c78225" exitCode=0 Oct 01 09:48:48 crc kubenswrapper[4787]: I1001 09:48:48.685258 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-wmqqd" Oct 01 09:48:48 crc kubenswrapper[4787]: I1001 09:48:48.685295 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-wmqqd" event={"ID":"be674cf7-d26e-464e-b661-4bff851d0413","Type":"ContainerDied","Data":"f84caac49271c19bcf1d172d60b7461e37f25120ee1add03510e9ee788c78225"} Oct 01 09:48:48 crc kubenswrapper[4787]: I1001 09:48:48.686198 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-wmqqd" event={"ID":"be674cf7-d26e-464e-b661-4bff851d0413","Type":"ContainerDied","Data":"0b72bec6e8b01ba95460ebfa652b73c4cd6c094f961ac7cf69f9a6b0e059471a"} Oct 01 09:48:48 crc kubenswrapper[4787]: I1001 09:48:48.686227 4787 scope.go:117] "RemoveContainer" containerID="f84caac49271c19bcf1d172d60b7461e37f25120ee1add03510e9ee788c78225" Oct 01 09:48:48 crc kubenswrapper[4787]: I1001 09:48:48.689797 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-jnbw4" event={"ID":"c6b4eeb7-99c2-44ff-9f71-e27c17c17649","Type":"ContainerStarted","Data":"883f6ea8a598fe89d21c1a4bef2750fd16105287cc0b90bc6f926b423f5c6fb8"} Oct 01 09:48:48 crc kubenswrapper[4787]: I1001 09:48:48.689822 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-jnbw4" event={"ID":"c6b4eeb7-99c2-44ff-9f71-e27c17c17649","Type":"ContainerStarted","Data":"f5abf28b9b2de221d2902b1e80bf30b84d3921fe47492440efb2bb69a7372a5d"} Oct 01 09:48:48 crc kubenswrapper[4787]: I1001 09:48:48.714469 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-jnbw4" podStartSLOduration=1.230012602 podStartE2EDuration="1.714444087s" podCreationTimestamp="2025-10-01 09:48:47 +0000 UTC" firstStartedPulling="2025-10-01 09:48:47.822771024 +0000 UTC m=+759.937915211" lastFinishedPulling="2025-10-01 09:48:48.307202509 +0000 UTC m=+760.422346696" observedRunningTime="2025-10-01 09:48:48.703874335 +0000 UTC m=+760.819018492" watchObservedRunningTime="2025-10-01 09:48:48.714444087 +0000 UTC m=+760.829588244" Oct 01 09:48:48 crc kubenswrapper[4787]: I1001 09:48:48.716516 4787 scope.go:117] "RemoveContainer" containerID="f84caac49271c19bcf1d172d60b7461e37f25120ee1add03510e9ee788c78225" Oct 01 09:48:48 crc kubenswrapper[4787]: E1001 09:48:48.721180 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f84caac49271c19bcf1d172d60b7461e37f25120ee1add03510e9ee788c78225\": container with ID starting with f84caac49271c19bcf1d172d60b7461e37f25120ee1add03510e9ee788c78225 not found: ID does not exist" containerID="f84caac49271c19bcf1d172d60b7461e37f25120ee1add03510e9ee788c78225" Oct 01 09:48:48 crc kubenswrapper[4787]: I1001 09:48:48.721245 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f84caac49271c19bcf1d172d60b7461e37f25120ee1add03510e9ee788c78225"} err="failed to get container status \"f84caac49271c19bcf1d172d60b7461e37f25120ee1add03510e9ee788c78225\": rpc error: code = NotFound desc = could not find container \"f84caac49271c19bcf1d172d60b7461e37f25120ee1add03510e9ee788c78225\": container with ID starting with f84caac49271c19bcf1d172d60b7461e37f25120ee1add03510e9ee788c78225 not found: ID does not exist" Oct 01 09:48:48 crc kubenswrapper[4787]: I1001 09:48:48.727639 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-wmqqd"] Oct 01 09:48:48 crc kubenswrapper[4787]: I1001 09:48:48.735779 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-wmqqd"] Oct 01 09:48:50 crc kubenswrapper[4787]: I1001 09:48:50.533702 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be674cf7-d26e-464e-b661-4bff851d0413" path="/var/lib/kubelet/pods/be674cf7-d26e-464e-b661-4bff851d0413/volumes" Oct 01 09:48:54 crc kubenswrapper[4787]: I1001 09:48:54.438034 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-p8sng"] Oct 01 09:48:54 crc kubenswrapper[4787]: E1001 09:48:54.438979 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be674cf7-d26e-464e-b661-4bff851d0413" containerName="registry-server" Oct 01 09:48:54 crc kubenswrapper[4787]: I1001 09:48:54.438997 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="be674cf7-d26e-464e-b661-4bff851d0413" containerName="registry-server" Oct 01 09:48:54 crc kubenswrapper[4787]: I1001 09:48:54.439162 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="be674cf7-d26e-464e-b661-4bff851d0413" containerName="registry-server" Oct 01 09:48:54 crc kubenswrapper[4787]: I1001 09:48:54.440367 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p8sng" Oct 01 09:48:54 crc kubenswrapper[4787]: I1001 09:48:54.447840 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p8sng"] Oct 01 09:48:54 crc kubenswrapper[4787]: I1001 09:48:54.518952 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwtr5\" (UniqueName: \"kubernetes.io/projected/eb54d02e-51d2-4068-a95c-3be6d561c05f-kube-api-access-gwtr5\") pod \"certified-operators-p8sng\" (UID: \"eb54d02e-51d2-4068-a95c-3be6d561c05f\") " pod="openshift-marketplace/certified-operators-p8sng" Oct 01 09:48:54 crc kubenswrapper[4787]: I1001 09:48:54.519015 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb54d02e-51d2-4068-a95c-3be6d561c05f-catalog-content\") pod \"certified-operators-p8sng\" (UID: \"eb54d02e-51d2-4068-a95c-3be6d561c05f\") " pod="openshift-marketplace/certified-operators-p8sng" Oct 01 09:48:54 crc kubenswrapper[4787]: I1001 09:48:54.519039 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb54d02e-51d2-4068-a95c-3be6d561c05f-utilities\") pod \"certified-operators-p8sng\" (UID: \"eb54d02e-51d2-4068-a95c-3be6d561c05f\") " pod="openshift-marketplace/certified-operators-p8sng" Oct 01 09:48:54 crc kubenswrapper[4787]: I1001 09:48:54.620975 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwtr5\" (UniqueName: \"kubernetes.io/projected/eb54d02e-51d2-4068-a95c-3be6d561c05f-kube-api-access-gwtr5\") pod \"certified-operators-p8sng\" (UID: \"eb54d02e-51d2-4068-a95c-3be6d561c05f\") " pod="openshift-marketplace/certified-operators-p8sng" Oct 01 09:48:54 crc kubenswrapper[4787]: I1001 09:48:54.621173 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb54d02e-51d2-4068-a95c-3be6d561c05f-catalog-content\") pod \"certified-operators-p8sng\" (UID: \"eb54d02e-51d2-4068-a95c-3be6d561c05f\") " pod="openshift-marketplace/certified-operators-p8sng" Oct 01 09:48:54 crc kubenswrapper[4787]: I1001 09:48:54.621204 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb54d02e-51d2-4068-a95c-3be6d561c05f-utilities\") pod \"certified-operators-p8sng\" (UID: \"eb54d02e-51d2-4068-a95c-3be6d561c05f\") " pod="openshift-marketplace/certified-operators-p8sng" Oct 01 09:48:54 crc kubenswrapper[4787]: I1001 09:48:54.621754 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb54d02e-51d2-4068-a95c-3be6d561c05f-catalog-content\") pod \"certified-operators-p8sng\" (UID: \"eb54d02e-51d2-4068-a95c-3be6d561c05f\") " pod="openshift-marketplace/certified-operators-p8sng" Oct 01 09:48:54 crc kubenswrapper[4787]: I1001 09:48:54.621876 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb54d02e-51d2-4068-a95c-3be6d561c05f-utilities\") pod \"certified-operators-p8sng\" (UID: \"eb54d02e-51d2-4068-a95c-3be6d561c05f\") " pod="openshift-marketplace/certified-operators-p8sng" Oct 01 09:48:54 crc kubenswrapper[4787]: I1001 09:48:54.643097 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwtr5\" (UniqueName: \"kubernetes.io/projected/eb54d02e-51d2-4068-a95c-3be6d561c05f-kube-api-access-gwtr5\") pod \"certified-operators-p8sng\" (UID: \"eb54d02e-51d2-4068-a95c-3be6d561c05f\") " pod="openshift-marketplace/certified-operators-p8sng" Oct 01 09:48:54 crc kubenswrapper[4787]: I1001 09:48:54.758677 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p8sng" Oct 01 09:48:55 crc kubenswrapper[4787]: I1001 09:48:55.250745 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p8sng"] Oct 01 09:48:55 crc kubenswrapper[4787]: I1001 09:48:55.759245 4787 generic.go:334] "Generic (PLEG): container finished" podID="eb54d02e-51d2-4068-a95c-3be6d561c05f" containerID="6401fdee424bf002578b6d49d9caa52d5ecf10465733219147e068d8a3a0972a" exitCode=0 Oct 01 09:48:55 crc kubenswrapper[4787]: I1001 09:48:55.759290 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p8sng" event={"ID":"eb54d02e-51d2-4068-a95c-3be6d561c05f","Type":"ContainerDied","Data":"6401fdee424bf002578b6d49d9caa52d5ecf10465733219147e068d8a3a0972a"} Oct 01 09:48:55 crc kubenswrapper[4787]: I1001 09:48:55.759315 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p8sng" event={"ID":"eb54d02e-51d2-4068-a95c-3be6d561c05f","Type":"ContainerStarted","Data":"78f5a1acace37de93f3ad0a2988eca8b5f6310ba95555641b45efc12b709bf2f"} Oct 01 09:48:55 crc kubenswrapper[4787]: I1001 09:48:55.826589 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dthpp"] Oct 01 09:48:55 crc kubenswrapper[4787]: I1001 09:48:55.828123 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dthpp" Oct 01 09:48:55 crc kubenswrapper[4787]: I1001 09:48:55.844024 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dthpp"] Oct 01 09:48:55 crc kubenswrapper[4787]: I1001 09:48:55.942031 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a96afbb-4648-480a-9ba8-cd36ca0d6962-catalog-content\") pod \"community-operators-dthpp\" (UID: \"3a96afbb-4648-480a-9ba8-cd36ca0d6962\") " pod="openshift-marketplace/community-operators-dthpp" Oct 01 09:48:55 crc kubenswrapper[4787]: I1001 09:48:55.942130 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a96afbb-4648-480a-9ba8-cd36ca0d6962-utilities\") pod \"community-operators-dthpp\" (UID: \"3a96afbb-4648-480a-9ba8-cd36ca0d6962\") " pod="openshift-marketplace/community-operators-dthpp" Oct 01 09:48:55 crc kubenswrapper[4787]: I1001 09:48:55.942225 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmszw\" (UniqueName: \"kubernetes.io/projected/3a96afbb-4648-480a-9ba8-cd36ca0d6962-kube-api-access-cmszw\") pod \"community-operators-dthpp\" (UID: \"3a96afbb-4648-480a-9ba8-cd36ca0d6962\") " pod="openshift-marketplace/community-operators-dthpp" Oct 01 09:48:56 crc kubenswrapper[4787]: I1001 09:48:56.043472 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a96afbb-4648-480a-9ba8-cd36ca0d6962-utilities\") pod \"community-operators-dthpp\" (UID: \"3a96afbb-4648-480a-9ba8-cd36ca0d6962\") " pod="openshift-marketplace/community-operators-dthpp" Oct 01 09:48:56 crc kubenswrapper[4787]: I1001 09:48:56.043567 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmszw\" (UniqueName: \"kubernetes.io/projected/3a96afbb-4648-480a-9ba8-cd36ca0d6962-kube-api-access-cmszw\") pod \"community-operators-dthpp\" (UID: \"3a96afbb-4648-480a-9ba8-cd36ca0d6962\") " pod="openshift-marketplace/community-operators-dthpp" Oct 01 09:48:56 crc kubenswrapper[4787]: I1001 09:48:56.043651 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a96afbb-4648-480a-9ba8-cd36ca0d6962-catalog-content\") pod \"community-operators-dthpp\" (UID: \"3a96afbb-4648-480a-9ba8-cd36ca0d6962\") " pod="openshift-marketplace/community-operators-dthpp" Oct 01 09:48:56 crc kubenswrapper[4787]: I1001 09:48:56.043885 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a96afbb-4648-480a-9ba8-cd36ca0d6962-utilities\") pod \"community-operators-dthpp\" (UID: \"3a96afbb-4648-480a-9ba8-cd36ca0d6962\") " pod="openshift-marketplace/community-operators-dthpp" Oct 01 09:48:56 crc kubenswrapper[4787]: I1001 09:48:56.044150 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a96afbb-4648-480a-9ba8-cd36ca0d6962-catalog-content\") pod \"community-operators-dthpp\" (UID: \"3a96afbb-4648-480a-9ba8-cd36ca0d6962\") " pod="openshift-marketplace/community-operators-dthpp" Oct 01 09:48:56 crc kubenswrapper[4787]: I1001 09:48:56.073921 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmszw\" (UniqueName: \"kubernetes.io/projected/3a96afbb-4648-480a-9ba8-cd36ca0d6962-kube-api-access-cmszw\") pod \"community-operators-dthpp\" (UID: \"3a96afbb-4648-480a-9ba8-cd36ca0d6962\") " pod="openshift-marketplace/community-operators-dthpp" Oct 01 09:48:56 crc kubenswrapper[4787]: I1001 09:48:56.149855 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dthpp" Oct 01 09:48:56 crc kubenswrapper[4787]: I1001 09:48:56.694522 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dthpp"] Oct 01 09:48:56 crc kubenswrapper[4787]: W1001 09:48:56.705846 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a96afbb_4648_480a_9ba8_cd36ca0d6962.slice/crio-03128c742e9c26ac42f9769f79893fdff7d5724536148c8776926b8887d1cd07 WatchSource:0}: Error finding container 03128c742e9c26ac42f9769f79893fdff7d5724536148c8776926b8887d1cd07: Status 404 returned error can't find the container with id 03128c742e9c26ac42f9769f79893fdff7d5724536148c8776926b8887d1cd07 Oct 01 09:48:56 crc kubenswrapper[4787]: I1001 09:48:56.769235 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dthpp" event={"ID":"3a96afbb-4648-480a-9ba8-cd36ca0d6962","Type":"ContainerStarted","Data":"03128c742e9c26ac42f9769f79893fdff7d5724536148c8776926b8887d1cd07"} Oct 01 09:48:56 crc kubenswrapper[4787]: I1001 09:48:56.773318 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p8sng" event={"ID":"eb54d02e-51d2-4068-a95c-3be6d561c05f","Type":"ContainerStarted","Data":"57fd157ff87803576437f5ef7073eb4c23a9b7300bb207669752662c83c82018"} Oct 01 09:48:57 crc kubenswrapper[4787]: I1001 09:48:57.358519 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-jnbw4" Oct 01 09:48:57 crc kubenswrapper[4787]: I1001 09:48:57.358559 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-jnbw4" Oct 01 09:48:57 crc kubenswrapper[4787]: I1001 09:48:57.396051 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-jnbw4" Oct 01 09:48:57 crc kubenswrapper[4787]: I1001 09:48:57.788545 4787 generic.go:334] "Generic (PLEG): container finished" podID="eb54d02e-51d2-4068-a95c-3be6d561c05f" containerID="57fd157ff87803576437f5ef7073eb4c23a9b7300bb207669752662c83c82018" exitCode=0 Oct 01 09:48:57 crc kubenswrapper[4787]: I1001 09:48:57.788816 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p8sng" event={"ID":"eb54d02e-51d2-4068-a95c-3be6d561c05f","Type":"ContainerDied","Data":"57fd157ff87803576437f5ef7073eb4c23a9b7300bb207669752662c83c82018"} Oct 01 09:48:57 crc kubenswrapper[4787]: I1001 09:48:57.794967 4787 generic.go:334] "Generic (PLEG): container finished" podID="3a96afbb-4648-480a-9ba8-cd36ca0d6962" containerID="9f99d26977251326995504a34e78099c4e327d8d8f2a0818f721d23377c899ad" exitCode=0 Oct 01 09:48:57 crc kubenswrapper[4787]: I1001 09:48:57.795633 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dthpp" event={"ID":"3a96afbb-4648-480a-9ba8-cd36ca0d6962","Type":"ContainerDied","Data":"9f99d26977251326995504a34e78099c4e327d8d8f2a0818f721d23377c899ad"} Oct 01 09:48:57 crc kubenswrapper[4787]: I1001 09:48:57.843835 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-jnbw4" Oct 01 09:48:58 crc kubenswrapper[4787]: I1001 09:48:58.804984 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dthpp" event={"ID":"3a96afbb-4648-480a-9ba8-cd36ca0d6962","Type":"ContainerStarted","Data":"991f7c19cc8c6314b2d33aaead78777d7762570d1a0fe7e3084caa07d28f3465"} Oct 01 09:48:58 crc kubenswrapper[4787]: I1001 09:48:58.809383 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p8sng" event={"ID":"eb54d02e-51d2-4068-a95c-3be6d561c05f","Type":"ContainerStarted","Data":"19447b46cc5d9d35de0e7500bef0824874238fdd53825403a238d63bb468b0cd"} Oct 01 09:48:58 crc kubenswrapper[4787]: I1001 09:48:58.869546 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-p8sng" podStartSLOduration=2.1794095430000002 podStartE2EDuration="4.86952222s" podCreationTimestamp="2025-10-01 09:48:54 +0000 UTC" firstStartedPulling="2025-10-01 09:48:55.760697499 +0000 UTC m=+767.875841656" lastFinishedPulling="2025-10-01 09:48:58.450810176 +0000 UTC m=+770.565954333" observedRunningTime="2025-10-01 09:48:58.865480636 +0000 UTC m=+770.980624803" watchObservedRunningTime="2025-10-01 09:48:58.86952222 +0000 UTC m=+770.984666377" Oct 01 09:48:58 crc kubenswrapper[4787]: I1001 09:48:58.920352 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh"] Oct 01 09:48:58 crc kubenswrapper[4787]: I1001 09:48:58.921772 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh" Oct 01 09:48:58 crc kubenswrapper[4787]: I1001 09:48:58.926247 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-8sp7m" Oct 01 09:48:58 crc kubenswrapper[4787]: I1001 09:48:58.942675 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh"] Oct 01 09:48:58 crc kubenswrapper[4787]: I1001 09:48:58.991134 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1-bundle\") pod \"0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh\" (UID: \"a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1\") " pod="openstack-operators/0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh" Oct 01 09:48:58 crc kubenswrapper[4787]: I1001 09:48:58.991224 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1-util\") pod \"0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh\" (UID: \"a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1\") " pod="openstack-operators/0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh" Oct 01 09:48:58 crc kubenswrapper[4787]: I1001 09:48:58.991276 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxxj7\" (UniqueName: \"kubernetes.io/projected/a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1-kube-api-access-qxxj7\") pod \"0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh\" (UID: \"a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1\") " pod="openstack-operators/0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh" Oct 01 09:48:59 crc kubenswrapper[4787]: I1001 09:48:59.093294 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1-bundle\") pod \"0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh\" (UID: \"a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1\") " pod="openstack-operators/0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh" Oct 01 09:48:59 crc kubenswrapper[4787]: I1001 09:48:59.093383 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1-util\") pod \"0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh\" (UID: \"a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1\") " pod="openstack-operators/0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh" Oct 01 09:48:59 crc kubenswrapper[4787]: I1001 09:48:59.093449 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxxj7\" (UniqueName: \"kubernetes.io/projected/a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1-kube-api-access-qxxj7\") pod \"0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh\" (UID: \"a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1\") " pod="openstack-operators/0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh" Oct 01 09:48:59 crc kubenswrapper[4787]: I1001 09:48:59.094107 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1-util\") pod \"0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh\" (UID: \"a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1\") " pod="openstack-operators/0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh" Oct 01 09:48:59 crc kubenswrapper[4787]: I1001 09:48:59.094395 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1-bundle\") pod \"0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh\" (UID: \"a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1\") " pod="openstack-operators/0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh" Oct 01 09:48:59 crc kubenswrapper[4787]: I1001 09:48:59.120116 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxxj7\" (UniqueName: \"kubernetes.io/projected/a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1-kube-api-access-qxxj7\") pod \"0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh\" (UID: \"a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1\") " pod="openstack-operators/0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh" Oct 01 09:48:59 crc kubenswrapper[4787]: I1001 09:48:59.239323 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh" Oct 01 09:48:59 crc kubenswrapper[4787]: I1001 09:48:59.750776 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh"] Oct 01 09:48:59 crc kubenswrapper[4787]: I1001 09:48:59.818202 4787 generic.go:334] "Generic (PLEG): container finished" podID="3a96afbb-4648-480a-9ba8-cd36ca0d6962" containerID="991f7c19cc8c6314b2d33aaead78777d7762570d1a0fe7e3084caa07d28f3465" exitCode=0 Oct 01 09:48:59 crc kubenswrapper[4787]: I1001 09:48:59.818381 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dthpp" event={"ID":"3a96afbb-4648-480a-9ba8-cd36ca0d6962","Type":"ContainerDied","Data":"991f7c19cc8c6314b2d33aaead78777d7762570d1a0fe7e3084caa07d28f3465"} Oct 01 09:48:59 crc kubenswrapper[4787]: I1001 09:48:59.819418 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh" event={"ID":"a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1","Type":"ContainerStarted","Data":"aa0887042fa266c21fd3b850610676c306e36f4b49e8b3e8a24dc66615bd08c9"} Oct 01 09:49:00 crc kubenswrapper[4787]: I1001 09:49:00.830116 4787 generic.go:334] "Generic (PLEG): container finished" podID="a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1" containerID="919d2f4ad23d5a697b80e58680e55ea2a6d885c8fd87a09c7959737a90eb7de2" exitCode=0 Oct 01 09:49:00 crc kubenswrapper[4787]: I1001 09:49:00.830200 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh" event={"ID":"a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1","Type":"ContainerDied","Data":"919d2f4ad23d5a697b80e58680e55ea2a6d885c8fd87a09c7959737a90eb7de2"} Oct 01 09:49:00 crc kubenswrapper[4787]: I1001 09:49:00.834441 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dthpp" event={"ID":"3a96afbb-4648-480a-9ba8-cd36ca0d6962","Type":"ContainerStarted","Data":"64c57e9cfc90108dc5f003744ae606e996e4e524078905cb14d9fd1d63f955a4"} Oct 01 09:49:00 crc kubenswrapper[4787]: I1001 09:49:00.870294 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dthpp" podStartSLOduration=3.448264611 podStartE2EDuration="5.87026161s" podCreationTimestamp="2025-10-01 09:48:55 +0000 UTC" firstStartedPulling="2025-10-01 09:48:57.800024661 +0000 UTC m=+769.915168818" lastFinishedPulling="2025-10-01 09:49:00.22202165 +0000 UTC m=+772.337165817" observedRunningTime="2025-10-01 09:49:00.869622553 +0000 UTC m=+772.984766740" watchObservedRunningTime="2025-10-01 09:49:00.87026161 +0000 UTC m=+772.985405797" Oct 01 09:49:03 crc kubenswrapper[4787]: I1001 09:49:03.872864 4787 generic.go:334] "Generic (PLEG): container finished" podID="a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1" containerID="a722816d5e1a888ec027bde13fd7364a912d34fb77d75086ebd5165b93ad4fba" exitCode=0 Oct 01 09:49:03 crc kubenswrapper[4787]: I1001 09:49:03.873012 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh" event={"ID":"a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1","Type":"ContainerDied","Data":"a722816d5e1a888ec027bde13fd7364a912d34fb77d75086ebd5165b93ad4fba"} Oct 01 09:49:04 crc kubenswrapper[4787]: I1001 09:49:04.759297 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-p8sng" Oct 01 09:49:04 crc kubenswrapper[4787]: I1001 09:49:04.759582 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-p8sng" Oct 01 09:49:04 crc kubenswrapper[4787]: I1001 09:49:04.813285 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-p8sng" Oct 01 09:49:04 crc kubenswrapper[4787]: I1001 09:49:04.884537 4787 generic.go:334] "Generic (PLEG): container finished" podID="a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1" containerID="af3e9e76896d0927d45dbd256bc920c5b7933d15bd37463544eaa8a57b71ced4" exitCode=0 Oct 01 09:49:04 crc kubenswrapper[4787]: I1001 09:49:04.884579 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh" event={"ID":"a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1","Type":"ContainerDied","Data":"af3e9e76896d0927d45dbd256bc920c5b7933d15bd37463544eaa8a57b71ced4"} Oct 01 09:49:04 crc kubenswrapper[4787]: I1001 09:49:04.928448 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-p8sng" Oct 01 09:49:05 crc kubenswrapper[4787]: I1001 09:49:05.832870 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5tmfr"] Oct 01 09:49:05 crc kubenswrapper[4787]: I1001 09:49:05.835464 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5tmfr" Oct 01 09:49:05 crc kubenswrapper[4787]: I1001 09:49:05.855825 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5tmfr"] Oct 01 09:49:05 crc kubenswrapper[4787]: I1001 09:49:05.895433 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c95df987-c9ac-4bc2-9077-30cb86f3fa7b-utilities\") pod \"redhat-operators-5tmfr\" (UID: \"c95df987-c9ac-4bc2-9077-30cb86f3fa7b\") " pod="openshift-marketplace/redhat-operators-5tmfr" Oct 01 09:49:05 crc kubenswrapper[4787]: I1001 09:49:05.896214 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c95df987-c9ac-4bc2-9077-30cb86f3fa7b-catalog-content\") pod \"redhat-operators-5tmfr\" (UID: \"c95df987-c9ac-4bc2-9077-30cb86f3fa7b\") " pod="openshift-marketplace/redhat-operators-5tmfr" Oct 01 09:49:05 crc kubenswrapper[4787]: I1001 09:49:05.896265 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44tmb\" (UniqueName: \"kubernetes.io/projected/c95df987-c9ac-4bc2-9077-30cb86f3fa7b-kube-api-access-44tmb\") pod \"redhat-operators-5tmfr\" (UID: \"c95df987-c9ac-4bc2-9077-30cb86f3fa7b\") " pod="openshift-marketplace/redhat-operators-5tmfr" Oct 01 09:49:05 crc kubenswrapper[4787]: I1001 09:49:05.997665 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c95df987-c9ac-4bc2-9077-30cb86f3fa7b-catalog-content\") pod \"redhat-operators-5tmfr\" (UID: \"c95df987-c9ac-4bc2-9077-30cb86f3fa7b\") " pod="openshift-marketplace/redhat-operators-5tmfr" Oct 01 09:49:05 crc kubenswrapper[4787]: I1001 09:49:05.998055 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44tmb\" (UniqueName: \"kubernetes.io/projected/c95df987-c9ac-4bc2-9077-30cb86f3fa7b-kube-api-access-44tmb\") pod \"redhat-operators-5tmfr\" (UID: \"c95df987-c9ac-4bc2-9077-30cb86f3fa7b\") " pod="openshift-marketplace/redhat-operators-5tmfr" Oct 01 09:49:05 crc kubenswrapper[4787]: I1001 09:49:05.998131 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c95df987-c9ac-4bc2-9077-30cb86f3fa7b-utilities\") pod \"redhat-operators-5tmfr\" (UID: \"c95df987-c9ac-4bc2-9077-30cb86f3fa7b\") " pod="openshift-marketplace/redhat-operators-5tmfr" Oct 01 09:49:05 crc kubenswrapper[4787]: I1001 09:49:05.999026 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c95df987-c9ac-4bc2-9077-30cb86f3fa7b-utilities\") pod \"redhat-operators-5tmfr\" (UID: \"c95df987-c9ac-4bc2-9077-30cb86f3fa7b\") " pod="openshift-marketplace/redhat-operators-5tmfr" Oct 01 09:49:05 crc kubenswrapper[4787]: I1001 09:49:05.999792 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c95df987-c9ac-4bc2-9077-30cb86f3fa7b-catalog-content\") pod \"redhat-operators-5tmfr\" (UID: \"c95df987-c9ac-4bc2-9077-30cb86f3fa7b\") " pod="openshift-marketplace/redhat-operators-5tmfr" Oct 01 09:49:06 crc kubenswrapper[4787]: I1001 09:49:06.030173 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44tmb\" (UniqueName: \"kubernetes.io/projected/c95df987-c9ac-4bc2-9077-30cb86f3fa7b-kube-api-access-44tmb\") pod \"redhat-operators-5tmfr\" (UID: \"c95df987-c9ac-4bc2-9077-30cb86f3fa7b\") " pod="openshift-marketplace/redhat-operators-5tmfr" Oct 01 09:49:06 crc kubenswrapper[4787]: I1001 09:49:06.153599 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dthpp" Oct 01 09:49:06 crc kubenswrapper[4787]: I1001 09:49:06.153664 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dthpp" Oct 01 09:49:06 crc kubenswrapper[4787]: I1001 09:49:06.155063 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5tmfr" Oct 01 09:49:06 crc kubenswrapper[4787]: I1001 09:49:06.208407 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dthpp" Oct 01 09:49:06 crc kubenswrapper[4787]: I1001 09:49:06.343852 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh" Oct 01 09:49:06 crc kubenswrapper[4787]: I1001 09:49:06.404750 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxxj7\" (UniqueName: \"kubernetes.io/projected/a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1-kube-api-access-qxxj7\") pod \"a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1\" (UID: \"a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1\") " Oct 01 09:49:06 crc kubenswrapper[4787]: I1001 09:49:06.404932 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1-bundle\") pod \"a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1\" (UID: \"a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1\") " Oct 01 09:49:06 crc kubenswrapper[4787]: I1001 09:49:06.406101 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1-bundle" (OuterVolumeSpecName: "bundle") pod "a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1" (UID: "a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:49:06 crc kubenswrapper[4787]: I1001 09:49:06.406214 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1-util\") pod \"a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1\" (UID: \"a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1\") " Oct 01 09:49:06 crc kubenswrapper[4787]: I1001 09:49:06.406978 4787 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:49:06 crc kubenswrapper[4787]: I1001 09:49:06.412679 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1-kube-api-access-qxxj7" (OuterVolumeSpecName: "kube-api-access-qxxj7") pod "a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1" (UID: "a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1"). InnerVolumeSpecName "kube-api-access-qxxj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:49:06 crc kubenswrapper[4787]: I1001 09:49:06.420628 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1-util" (OuterVolumeSpecName: "util") pod "a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1" (UID: "a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:49:06 crc kubenswrapper[4787]: I1001 09:49:06.509154 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxxj7\" (UniqueName: \"kubernetes.io/projected/a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1-kube-api-access-qxxj7\") on node \"crc\" DevicePath \"\"" Oct 01 09:49:06 crc kubenswrapper[4787]: I1001 09:49:06.509208 4787 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1-util\") on node \"crc\" DevicePath \"\"" Oct 01 09:49:06 crc kubenswrapper[4787]: I1001 09:49:06.648198 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5tmfr"] Oct 01 09:49:06 crc kubenswrapper[4787]: W1001 09:49:06.655833 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc95df987_c9ac_4bc2_9077_30cb86f3fa7b.slice/crio-d983ca32a5a6e0bef7a14942bc2d85eaeec06a586df6821edafc001c5c4968a0 WatchSource:0}: Error finding container d983ca32a5a6e0bef7a14942bc2d85eaeec06a586df6821edafc001c5c4968a0: Status 404 returned error can't find the container with id d983ca32a5a6e0bef7a14942bc2d85eaeec06a586df6821edafc001c5c4968a0 Oct 01 09:49:06 crc kubenswrapper[4787]: I1001 09:49:06.815486 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p8sng"] Oct 01 09:49:06 crc kubenswrapper[4787]: I1001 09:49:06.899772 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh" Oct 01 09:49:06 crc kubenswrapper[4787]: I1001 09:49:06.899763 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh" event={"ID":"a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1","Type":"ContainerDied","Data":"aa0887042fa266c21fd3b850610676c306e36f4b49e8b3e8a24dc66615bd08c9"} Oct 01 09:49:06 crc kubenswrapper[4787]: I1001 09:49:06.899895 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa0887042fa266c21fd3b850610676c306e36f4b49e8b3e8a24dc66615bd08c9" Oct 01 09:49:06 crc kubenswrapper[4787]: I1001 09:49:06.901341 4787 generic.go:334] "Generic (PLEG): container finished" podID="c95df987-c9ac-4bc2-9077-30cb86f3fa7b" containerID="3f46cd8dcfae5b66f12dd446fd8722d0470a69575d836308c0c93f59bc392ec6" exitCode=0 Oct 01 09:49:06 crc kubenswrapper[4787]: I1001 09:49:06.901448 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5tmfr" event={"ID":"c95df987-c9ac-4bc2-9077-30cb86f3fa7b","Type":"ContainerDied","Data":"3f46cd8dcfae5b66f12dd446fd8722d0470a69575d836308c0c93f59bc392ec6"} Oct 01 09:49:06 crc kubenswrapper[4787]: I1001 09:49:06.901879 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5tmfr" event={"ID":"c95df987-c9ac-4bc2-9077-30cb86f3fa7b","Type":"ContainerStarted","Data":"d983ca32a5a6e0bef7a14942bc2d85eaeec06a586df6821edafc001c5c4968a0"} Oct 01 09:49:06 crc kubenswrapper[4787]: I1001 09:49:06.902063 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-p8sng" podUID="eb54d02e-51d2-4068-a95c-3be6d561c05f" containerName="registry-server" containerID="cri-o://19447b46cc5d9d35de0e7500bef0824874238fdd53825403a238d63bb468b0cd" gracePeriod=2 Oct 01 09:49:06 crc kubenswrapper[4787]: I1001 09:49:06.987393 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dthpp" Oct 01 09:49:07 crc kubenswrapper[4787]: I1001 09:49:07.387053 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p8sng" Oct 01 09:49:07 crc kubenswrapper[4787]: I1001 09:49:07.424101 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gwtr5\" (UniqueName: \"kubernetes.io/projected/eb54d02e-51d2-4068-a95c-3be6d561c05f-kube-api-access-gwtr5\") pod \"eb54d02e-51d2-4068-a95c-3be6d561c05f\" (UID: \"eb54d02e-51d2-4068-a95c-3be6d561c05f\") " Oct 01 09:49:07 crc kubenswrapper[4787]: I1001 09:49:07.424165 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb54d02e-51d2-4068-a95c-3be6d561c05f-utilities\") pod \"eb54d02e-51d2-4068-a95c-3be6d561c05f\" (UID: \"eb54d02e-51d2-4068-a95c-3be6d561c05f\") " Oct 01 09:49:07 crc kubenswrapper[4787]: I1001 09:49:07.424261 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb54d02e-51d2-4068-a95c-3be6d561c05f-catalog-content\") pod \"eb54d02e-51d2-4068-a95c-3be6d561c05f\" (UID: \"eb54d02e-51d2-4068-a95c-3be6d561c05f\") " Oct 01 09:49:07 crc kubenswrapper[4787]: I1001 09:49:07.425186 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb54d02e-51d2-4068-a95c-3be6d561c05f-utilities" (OuterVolumeSpecName: "utilities") pod "eb54d02e-51d2-4068-a95c-3be6d561c05f" (UID: "eb54d02e-51d2-4068-a95c-3be6d561c05f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:49:07 crc kubenswrapper[4787]: I1001 09:49:07.431093 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb54d02e-51d2-4068-a95c-3be6d561c05f-kube-api-access-gwtr5" (OuterVolumeSpecName: "kube-api-access-gwtr5") pod "eb54d02e-51d2-4068-a95c-3be6d561c05f" (UID: "eb54d02e-51d2-4068-a95c-3be6d561c05f"). InnerVolumeSpecName "kube-api-access-gwtr5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:49:07 crc kubenswrapper[4787]: I1001 09:49:07.468060 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb54d02e-51d2-4068-a95c-3be6d561c05f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eb54d02e-51d2-4068-a95c-3be6d561c05f" (UID: "eb54d02e-51d2-4068-a95c-3be6d561c05f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:49:07 crc kubenswrapper[4787]: I1001 09:49:07.526268 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gwtr5\" (UniqueName: \"kubernetes.io/projected/eb54d02e-51d2-4068-a95c-3be6d561c05f-kube-api-access-gwtr5\") on node \"crc\" DevicePath \"\"" Oct 01 09:49:07 crc kubenswrapper[4787]: I1001 09:49:07.526316 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb54d02e-51d2-4068-a95c-3be6d561c05f-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:49:07 crc kubenswrapper[4787]: I1001 09:49:07.526334 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb54d02e-51d2-4068-a95c-3be6d561c05f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:49:07 crc kubenswrapper[4787]: I1001 09:49:07.915440 4787 generic.go:334] "Generic (PLEG): container finished" podID="eb54d02e-51d2-4068-a95c-3be6d561c05f" containerID="19447b46cc5d9d35de0e7500bef0824874238fdd53825403a238d63bb468b0cd" exitCode=0 Oct 01 09:49:07 crc kubenswrapper[4787]: I1001 09:49:07.915520 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p8sng" event={"ID":"eb54d02e-51d2-4068-a95c-3be6d561c05f","Type":"ContainerDied","Data":"19447b46cc5d9d35de0e7500bef0824874238fdd53825403a238d63bb468b0cd"} Oct 01 09:49:07 crc kubenswrapper[4787]: I1001 09:49:07.915954 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p8sng" event={"ID":"eb54d02e-51d2-4068-a95c-3be6d561c05f","Type":"ContainerDied","Data":"78f5a1acace37de93f3ad0a2988eca8b5f6310ba95555641b45efc12b709bf2f"} Oct 01 09:49:07 crc kubenswrapper[4787]: I1001 09:49:07.916017 4787 scope.go:117] "RemoveContainer" containerID="19447b46cc5d9d35de0e7500bef0824874238fdd53825403a238d63bb468b0cd" Oct 01 09:49:07 crc kubenswrapper[4787]: I1001 09:49:07.915602 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p8sng" Oct 01 09:49:07 crc kubenswrapper[4787]: I1001 09:49:07.923067 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5tmfr" event={"ID":"c95df987-c9ac-4bc2-9077-30cb86f3fa7b","Type":"ContainerStarted","Data":"9bbfbb7da6ab1c8609b2283243e1cb185a445a698d7cace7a4b479db6809fc15"} Oct 01 09:49:07 crc kubenswrapper[4787]: I1001 09:49:07.944399 4787 scope.go:117] "RemoveContainer" containerID="57fd157ff87803576437f5ef7073eb4c23a9b7300bb207669752662c83c82018" Oct 01 09:49:07 crc kubenswrapper[4787]: I1001 09:49:07.967655 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p8sng"] Oct 01 09:49:07 crc kubenswrapper[4787]: I1001 09:49:07.971195 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-p8sng"] Oct 01 09:49:07 crc kubenswrapper[4787]: I1001 09:49:07.978873 4787 scope.go:117] "RemoveContainer" containerID="6401fdee424bf002578b6d49d9caa52d5ecf10465733219147e068d8a3a0972a" Oct 01 09:49:07 crc kubenswrapper[4787]: I1001 09:49:07.999554 4787 scope.go:117] "RemoveContainer" containerID="19447b46cc5d9d35de0e7500bef0824874238fdd53825403a238d63bb468b0cd" Oct 01 09:49:08 crc kubenswrapper[4787]: E1001 09:49:08.000602 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19447b46cc5d9d35de0e7500bef0824874238fdd53825403a238d63bb468b0cd\": container with ID starting with 19447b46cc5d9d35de0e7500bef0824874238fdd53825403a238d63bb468b0cd not found: ID does not exist" containerID="19447b46cc5d9d35de0e7500bef0824874238fdd53825403a238d63bb468b0cd" Oct 01 09:49:08 crc kubenswrapper[4787]: I1001 09:49:08.000663 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19447b46cc5d9d35de0e7500bef0824874238fdd53825403a238d63bb468b0cd"} err="failed to get container status \"19447b46cc5d9d35de0e7500bef0824874238fdd53825403a238d63bb468b0cd\": rpc error: code = NotFound desc = could not find container \"19447b46cc5d9d35de0e7500bef0824874238fdd53825403a238d63bb468b0cd\": container with ID starting with 19447b46cc5d9d35de0e7500bef0824874238fdd53825403a238d63bb468b0cd not found: ID does not exist" Oct 01 09:49:08 crc kubenswrapper[4787]: I1001 09:49:08.000697 4787 scope.go:117] "RemoveContainer" containerID="57fd157ff87803576437f5ef7073eb4c23a9b7300bb207669752662c83c82018" Oct 01 09:49:08 crc kubenswrapper[4787]: E1001 09:49:08.003583 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57fd157ff87803576437f5ef7073eb4c23a9b7300bb207669752662c83c82018\": container with ID starting with 57fd157ff87803576437f5ef7073eb4c23a9b7300bb207669752662c83c82018 not found: ID does not exist" containerID="57fd157ff87803576437f5ef7073eb4c23a9b7300bb207669752662c83c82018" Oct 01 09:49:08 crc kubenswrapper[4787]: I1001 09:49:08.003665 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57fd157ff87803576437f5ef7073eb4c23a9b7300bb207669752662c83c82018"} err="failed to get container status \"57fd157ff87803576437f5ef7073eb4c23a9b7300bb207669752662c83c82018\": rpc error: code = NotFound desc = could not find container \"57fd157ff87803576437f5ef7073eb4c23a9b7300bb207669752662c83c82018\": container with ID starting with 57fd157ff87803576437f5ef7073eb4c23a9b7300bb207669752662c83c82018 not found: ID does not exist" Oct 01 09:49:08 crc kubenswrapper[4787]: I1001 09:49:08.003711 4787 scope.go:117] "RemoveContainer" containerID="6401fdee424bf002578b6d49d9caa52d5ecf10465733219147e068d8a3a0972a" Oct 01 09:49:08 crc kubenswrapper[4787]: E1001 09:49:08.005540 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6401fdee424bf002578b6d49d9caa52d5ecf10465733219147e068d8a3a0972a\": container with ID starting with 6401fdee424bf002578b6d49d9caa52d5ecf10465733219147e068d8a3a0972a not found: ID does not exist" containerID="6401fdee424bf002578b6d49d9caa52d5ecf10465733219147e068d8a3a0972a" Oct 01 09:49:08 crc kubenswrapper[4787]: I1001 09:49:08.005624 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6401fdee424bf002578b6d49d9caa52d5ecf10465733219147e068d8a3a0972a"} err="failed to get container status \"6401fdee424bf002578b6d49d9caa52d5ecf10465733219147e068d8a3a0972a\": rpc error: code = NotFound desc = could not find container \"6401fdee424bf002578b6d49d9caa52d5ecf10465733219147e068d8a3a0972a\": container with ID starting with 6401fdee424bf002578b6d49d9caa52d5ecf10465733219147e068d8a3a0972a not found: ID does not exist" Oct 01 09:49:08 crc kubenswrapper[4787]: I1001 09:49:08.535770 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb54d02e-51d2-4068-a95c-3be6d561c05f" path="/var/lib/kubelet/pods/eb54d02e-51d2-4068-a95c-3be6d561c05f/volumes" Oct 01 09:49:08 crc kubenswrapper[4787]: I1001 09:49:08.933176 4787 generic.go:334] "Generic (PLEG): container finished" podID="c95df987-c9ac-4bc2-9077-30cb86f3fa7b" containerID="9bbfbb7da6ab1c8609b2283243e1cb185a445a698d7cace7a4b479db6809fc15" exitCode=0 Oct 01 09:49:08 crc kubenswrapper[4787]: I1001 09:49:08.933275 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5tmfr" event={"ID":"c95df987-c9ac-4bc2-9077-30cb86f3fa7b","Type":"ContainerDied","Data":"9bbfbb7da6ab1c8609b2283243e1cb185a445a698d7cace7a4b479db6809fc15"} Oct 01 09:49:09 crc kubenswrapper[4787]: I1001 09:49:09.414926 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dthpp"] Oct 01 09:49:09 crc kubenswrapper[4787]: I1001 09:49:09.415304 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-dthpp" podUID="3a96afbb-4648-480a-9ba8-cd36ca0d6962" containerName="registry-server" containerID="cri-o://64c57e9cfc90108dc5f003744ae606e996e4e524078905cb14d9fd1d63f955a4" gracePeriod=2 Oct 01 09:49:09 crc kubenswrapper[4787]: I1001 09:49:09.881807 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dthpp" Oct 01 09:49:09 crc kubenswrapper[4787]: I1001 09:49:09.941654 4787 generic.go:334] "Generic (PLEG): container finished" podID="3a96afbb-4648-480a-9ba8-cd36ca0d6962" containerID="64c57e9cfc90108dc5f003744ae606e996e4e524078905cb14d9fd1d63f955a4" exitCode=0 Oct 01 09:49:09 crc kubenswrapper[4787]: I1001 09:49:09.941751 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dthpp" event={"ID":"3a96afbb-4648-480a-9ba8-cd36ca0d6962","Type":"ContainerDied","Data":"64c57e9cfc90108dc5f003744ae606e996e4e524078905cb14d9fd1d63f955a4"} Oct 01 09:49:09 crc kubenswrapper[4787]: I1001 09:49:09.942500 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dthpp" event={"ID":"3a96afbb-4648-480a-9ba8-cd36ca0d6962","Type":"ContainerDied","Data":"03128c742e9c26ac42f9769f79893fdff7d5724536148c8776926b8887d1cd07"} Oct 01 09:49:09 crc kubenswrapper[4787]: I1001 09:49:09.942524 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dthpp" Oct 01 09:49:09 crc kubenswrapper[4787]: I1001 09:49:09.942535 4787 scope.go:117] "RemoveContainer" containerID="64c57e9cfc90108dc5f003744ae606e996e4e524078905cb14d9fd1d63f955a4" Oct 01 09:49:09 crc kubenswrapper[4787]: I1001 09:49:09.956835 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5tmfr" event={"ID":"c95df987-c9ac-4bc2-9077-30cb86f3fa7b","Type":"ContainerStarted","Data":"ce11f464aa1b3fe71d81c2de46eb18219d477d4cc54163d445e2734353941a35"} Oct 01 09:49:09 crc kubenswrapper[4787]: I1001 09:49:09.965883 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a96afbb-4648-480a-9ba8-cd36ca0d6962-utilities\") pod \"3a96afbb-4648-480a-9ba8-cd36ca0d6962\" (UID: \"3a96afbb-4648-480a-9ba8-cd36ca0d6962\") " Oct 01 09:49:09 crc kubenswrapper[4787]: I1001 09:49:09.966136 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmszw\" (UniqueName: \"kubernetes.io/projected/3a96afbb-4648-480a-9ba8-cd36ca0d6962-kube-api-access-cmszw\") pod \"3a96afbb-4648-480a-9ba8-cd36ca0d6962\" (UID: \"3a96afbb-4648-480a-9ba8-cd36ca0d6962\") " Oct 01 09:49:09 crc kubenswrapper[4787]: I1001 09:49:09.966699 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a96afbb-4648-480a-9ba8-cd36ca0d6962-catalog-content\") pod \"3a96afbb-4648-480a-9ba8-cd36ca0d6962\" (UID: \"3a96afbb-4648-480a-9ba8-cd36ca0d6962\") " Oct 01 09:49:09 crc kubenswrapper[4787]: I1001 09:49:09.967354 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a96afbb-4648-480a-9ba8-cd36ca0d6962-utilities" (OuterVolumeSpecName: "utilities") pod "3a96afbb-4648-480a-9ba8-cd36ca0d6962" (UID: "3a96afbb-4648-480a-9ba8-cd36ca0d6962"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:49:10 crc kubenswrapper[4787]: I1001 09:49:10.000332 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a96afbb-4648-480a-9ba8-cd36ca0d6962-kube-api-access-cmszw" (OuterVolumeSpecName: "kube-api-access-cmszw") pod "3a96afbb-4648-480a-9ba8-cd36ca0d6962" (UID: "3a96afbb-4648-480a-9ba8-cd36ca0d6962"). InnerVolumeSpecName "kube-api-access-cmszw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:49:10 crc kubenswrapper[4787]: I1001 09:49:10.000403 4787 scope.go:117] "RemoveContainer" containerID="991f7c19cc8c6314b2d33aaead78777d7762570d1a0fe7e3084caa07d28f3465" Oct 01 09:49:10 crc kubenswrapper[4787]: I1001 09:49:10.059473 4787 scope.go:117] "RemoveContainer" containerID="9f99d26977251326995504a34e78099c4e327d8d8f2a0818f721d23377c899ad" Oct 01 09:49:10 crc kubenswrapper[4787]: I1001 09:49:10.069397 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a96afbb-4648-480a-9ba8-cd36ca0d6962-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:49:10 crc kubenswrapper[4787]: I1001 09:49:10.069433 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmszw\" (UniqueName: \"kubernetes.io/projected/3a96afbb-4648-480a-9ba8-cd36ca0d6962-kube-api-access-cmszw\") on node \"crc\" DevicePath \"\"" Oct 01 09:49:10 crc kubenswrapper[4787]: I1001 09:49:10.129292 4787 scope.go:117] "RemoveContainer" containerID="64c57e9cfc90108dc5f003744ae606e996e4e524078905cb14d9fd1d63f955a4" Oct 01 09:49:10 crc kubenswrapper[4787]: E1001 09:49:10.132271 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64c57e9cfc90108dc5f003744ae606e996e4e524078905cb14d9fd1d63f955a4\": container with ID starting with 64c57e9cfc90108dc5f003744ae606e996e4e524078905cb14d9fd1d63f955a4 not found: ID does not exist" containerID="64c57e9cfc90108dc5f003744ae606e996e4e524078905cb14d9fd1d63f955a4" Oct 01 09:49:10 crc kubenswrapper[4787]: I1001 09:49:10.132315 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64c57e9cfc90108dc5f003744ae606e996e4e524078905cb14d9fd1d63f955a4"} err="failed to get container status \"64c57e9cfc90108dc5f003744ae606e996e4e524078905cb14d9fd1d63f955a4\": rpc error: code = NotFound desc = could not find container \"64c57e9cfc90108dc5f003744ae606e996e4e524078905cb14d9fd1d63f955a4\": container with ID starting with 64c57e9cfc90108dc5f003744ae606e996e4e524078905cb14d9fd1d63f955a4 not found: ID does not exist" Oct 01 09:49:10 crc kubenswrapper[4787]: I1001 09:49:10.132342 4787 scope.go:117] "RemoveContainer" containerID="991f7c19cc8c6314b2d33aaead78777d7762570d1a0fe7e3084caa07d28f3465" Oct 01 09:49:10 crc kubenswrapper[4787]: E1001 09:49:10.136184 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"991f7c19cc8c6314b2d33aaead78777d7762570d1a0fe7e3084caa07d28f3465\": container with ID starting with 991f7c19cc8c6314b2d33aaead78777d7762570d1a0fe7e3084caa07d28f3465 not found: ID does not exist" containerID="991f7c19cc8c6314b2d33aaead78777d7762570d1a0fe7e3084caa07d28f3465" Oct 01 09:49:10 crc kubenswrapper[4787]: I1001 09:49:10.136226 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"991f7c19cc8c6314b2d33aaead78777d7762570d1a0fe7e3084caa07d28f3465"} err="failed to get container status \"991f7c19cc8c6314b2d33aaead78777d7762570d1a0fe7e3084caa07d28f3465\": rpc error: code = NotFound desc = could not find container \"991f7c19cc8c6314b2d33aaead78777d7762570d1a0fe7e3084caa07d28f3465\": container with ID starting with 991f7c19cc8c6314b2d33aaead78777d7762570d1a0fe7e3084caa07d28f3465 not found: ID does not exist" Oct 01 09:49:10 crc kubenswrapper[4787]: I1001 09:49:10.136250 4787 scope.go:117] "RemoveContainer" containerID="9f99d26977251326995504a34e78099c4e327d8d8f2a0818f721d23377c899ad" Oct 01 09:49:10 crc kubenswrapper[4787]: E1001 09:49:10.143252 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f99d26977251326995504a34e78099c4e327d8d8f2a0818f721d23377c899ad\": container with ID starting with 9f99d26977251326995504a34e78099c4e327d8d8f2a0818f721d23377c899ad not found: ID does not exist" containerID="9f99d26977251326995504a34e78099c4e327d8d8f2a0818f721d23377c899ad" Oct 01 09:49:10 crc kubenswrapper[4787]: I1001 09:49:10.143305 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f99d26977251326995504a34e78099c4e327d8d8f2a0818f721d23377c899ad"} err="failed to get container status \"9f99d26977251326995504a34e78099c4e327d8d8f2a0818f721d23377c899ad\": rpc error: code = NotFound desc = could not find container \"9f99d26977251326995504a34e78099c4e327d8d8f2a0818f721d23377c899ad\": container with ID starting with 9f99d26977251326995504a34e78099c4e327d8d8f2a0818f721d23377c899ad not found: ID does not exist" Oct 01 09:49:10 crc kubenswrapper[4787]: I1001 09:49:10.348213 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5tmfr" podStartSLOduration=2.843246526 podStartE2EDuration="5.348194409s" podCreationTimestamp="2025-10-01 09:49:05 +0000 UTC" firstStartedPulling="2025-10-01 09:49:06.903494236 +0000 UTC m=+779.018638393" lastFinishedPulling="2025-10-01 09:49:09.408442119 +0000 UTC m=+781.523586276" observedRunningTime="2025-10-01 09:49:10.012160392 +0000 UTC m=+782.127304569" watchObservedRunningTime="2025-10-01 09:49:10.348194409 +0000 UTC m=+782.463338566" Oct 01 09:49:10 crc kubenswrapper[4787]: I1001 09:49:10.350100 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-676c66f88b-7hgdc"] Oct 01 09:49:10 crc kubenswrapper[4787]: E1001 09:49:10.350316 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1" containerName="extract" Oct 01 09:49:10 crc kubenswrapper[4787]: I1001 09:49:10.350334 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1" containerName="extract" Oct 01 09:49:10 crc kubenswrapper[4787]: E1001 09:49:10.350346 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb54d02e-51d2-4068-a95c-3be6d561c05f" containerName="extract-utilities" Oct 01 09:49:10 crc kubenswrapper[4787]: I1001 09:49:10.350354 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb54d02e-51d2-4068-a95c-3be6d561c05f" containerName="extract-utilities" Oct 01 09:49:10 crc kubenswrapper[4787]: E1001 09:49:10.350363 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb54d02e-51d2-4068-a95c-3be6d561c05f" containerName="registry-server" Oct 01 09:49:10 crc kubenswrapper[4787]: I1001 09:49:10.350371 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb54d02e-51d2-4068-a95c-3be6d561c05f" containerName="registry-server" Oct 01 09:49:10 crc kubenswrapper[4787]: E1001 09:49:10.350382 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb54d02e-51d2-4068-a95c-3be6d561c05f" containerName="extract-content" Oct 01 09:49:10 crc kubenswrapper[4787]: I1001 09:49:10.350389 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb54d02e-51d2-4068-a95c-3be6d561c05f" containerName="extract-content" Oct 01 09:49:10 crc kubenswrapper[4787]: E1001 09:49:10.350399 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a96afbb-4648-480a-9ba8-cd36ca0d6962" containerName="extract-content" Oct 01 09:49:10 crc kubenswrapper[4787]: I1001 09:49:10.350404 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a96afbb-4648-480a-9ba8-cd36ca0d6962" containerName="extract-content" Oct 01 09:49:10 crc kubenswrapper[4787]: E1001 09:49:10.350419 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a96afbb-4648-480a-9ba8-cd36ca0d6962" containerName="registry-server" Oct 01 09:49:10 crc kubenswrapper[4787]: I1001 09:49:10.350425 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a96afbb-4648-480a-9ba8-cd36ca0d6962" containerName="registry-server" Oct 01 09:49:10 crc kubenswrapper[4787]: E1001 09:49:10.350434 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a96afbb-4648-480a-9ba8-cd36ca0d6962" containerName="extract-utilities" Oct 01 09:49:10 crc kubenswrapper[4787]: I1001 09:49:10.350441 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a96afbb-4648-480a-9ba8-cd36ca0d6962" containerName="extract-utilities" Oct 01 09:49:10 crc kubenswrapper[4787]: E1001 09:49:10.350451 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1" containerName="pull" Oct 01 09:49:10 crc kubenswrapper[4787]: I1001 09:49:10.350457 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1" containerName="pull" Oct 01 09:49:10 crc kubenswrapper[4787]: E1001 09:49:10.350465 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1" containerName="util" Oct 01 09:49:10 crc kubenswrapper[4787]: I1001 09:49:10.350470 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1" containerName="util" Oct 01 09:49:10 crc kubenswrapper[4787]: I1001 09:49:10.350571 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb54d02e-51d2-4068-a95c-3be6d561c05f" containerName="registry-server" Oct 01 09:49:10 crc kubenswrapper[4787]: I1001 09:49:10.350583 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1" containerName="extract" Oct 01 09:49:10 crc kubenswrapper[4787]: I1001 09:49:10.350594 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a96afbb-4648-480a-9ba8-cd36ca0d6962" containerName="registry-server" Oct 01 09:49:10 crc kubenswrapper[4787]: I1001 09:49:10.351222 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-676c66f88b-7hgdc" Oct 01 09:49:10 crc kubenswrapper[4787]: I1001 09:49:10.353374 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-5qh52" Oct 01 09:49:10 crc kubenswrapper[4787]: I1001 09:49:10.377278 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-676c66f88b-7hgdc"] Oct 01 09:49:10 crc kubenswrapper[4787]: I1001 09:49:10.387613 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a96afbb-4648-480a-9ba8-cd36ca0d6962-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3a96afbb-4648-480a-9ba8-cd36ca0d6962" (UID: "3a96afbb-4648-480a-9ba8-cd36ca0d6962"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:49:10 crc kubenswrapper[4787]: I1001 09:49:10.474583 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2f87j\" (UniqueName: \"kubernetes.io/projected/faaa86ca-d16f-4ee1-9a9a-15c54aca2a8b-kube-api-access-2f87j\") pod \"openstack-operator-controller-operator-676c66f88b-7hgdc\" (UID: \"faaa86ca-d16f-4ee1-9a9a-15c54aca2a8b\") " pod="openstack-operators/openstack-operator-controller-operator-676c66f88b-7hgdc" Oct 01 09:49:10 crc kubenswrapper[4787]: I1001 09:49:10.474738 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a96afbb-4648-480a-9ba8-cd36ca0d6962-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:49:10 crc kubenswrapper[4787]: I1001 09:49:10.561893 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dthpp"] Oct 01 09:49:10 crc kubenswrapper[4787]: I1001 09:49:10.566272 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-dthpp"] Oct 01 09:49:10 crc kubenswrapper[4787]: I1001 09:49:10.576049 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2f87j\" (UniqueName: \"kubernetes.io/projected/faaa86ca-d16f-4ee1-9a9a-15c54aca2a8b-kube-api-access-2f87j\") pod \"openstack-operator-controller-operator-676c66f88b-7hgdc\" (UID: \"faaa86ca-d16f-4ee1-9a9a-15c54aca2a8b\") " pod="openstack-operators/openstack-operator-controller-operator-676c66f88b-7hgdc" Oct 01 09:49:10 crc kubenswrapper[4787]: I1001 09:49:10.602144 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2f87j\" (UniqueName: \"kubernetes.io/projected/faaa86ca-d16f-4ee1-9a9a-15c54aca2a8b-kube-api-access-2f87j\") pod \"openstack-operator-controller-operator-676c66f88b-7hgdc\" (UID: \"faaa86ca-d16f-4ee1-9a9a-15c54aca2a8b\") " pod="openstack-operators/openstack-operator-controller-operator-676c66f88b-7hgdc" Oct 01 09:49:10 crc kubenswrapper[4787]: I1001 09:49:10.669315 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-676c66f88b-7hgdc" Oct 01 09:49:11 crc kubenswrapper[4787]: I1001 09:49:11.164295 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-676c66f88b-7hgdc"] Oct 01 09:49:11 crc kubenswrapper[4787]: I1001 09:49:11.982255 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-676c66f88b-7hgdc" event={"ID":"faaa86ca-d16f-4ee1-9a9a-15c54aca2a8b","Type":"ContainerStarted","Data":"e4195dc25e81be0f78edd76aff934f210caf9693d82bfcdc2abc1c675f59bea5"} Oct 01 09:49:12 crc kubenswrapper[4787]: I1001 09:49:12.532770 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a96afbb-4648-480a-9ba8-cd36ca0d6962" path="/var/lib/kubelet/pods/3a96afbb-4648-480a-9ba8-cd36ca0d6962/volumes" Oct 01 09:49:16 crc kubenswrapper[4787]: I1001 09:49:16.015420 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-676c66f88b-7hgdc" event={"ID":"faaa86ca-d16f-4ee1-9a9a-15c54aca2a8b","Type":"ContainerStarted","Data":"51934f6600b07c14bb5ef26b08d0a11c14b0f15fa910987388464dd4100c68a6"} Oct 01 09:49:16 crc kubenswrapper[4787]: I1001 09:49:16.155582 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5tmfr" Oct 01 09:49:16 crc kubenswrapper[4787]: I1001 09:49:16.155636 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5tmfr" Oct 01 09:49:16 crc kubenswrapper[4787]: I1001 09:49:16.200529 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5tmfr" Oct 01 09:49:17 crc kubenswrapper[4787]: I1001 09:49:17.065395 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5tmfr" Oct 01 09:49:18 crc kubenswrapper[4787]: I1001 09:49:18.027867 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-676c66f88b-7hgdc" event={"ID":"faaa86ca-d16f-4ee1-9a9a-15c54aca2a8b","Type":"ContainerStarted","Data":"8944c5fe02838df4b258695dacb66b4165c941cb961a626b99d28db4633cb447"} Oct 01 09:49:18 crc kubenswrapper[4787]: I1001 09:49:18.028455 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-676c66f88b-7hgdc" Oct 01 09:49:18 crc kubenswrapper[4787]: I1001 09:49:18.059723 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-676c66f88b-7hgdc" podStartSLOduration=1.5306533340000001 podStartE2EDuration="8.059706048s" podCreationTimestamp="2025-10-01 09:49:10 +0000 UTC" firstStartedPulling="2025-10-01 09:49:11.176927852 +0000 UTC m=+783.292072009" lastFinishedPulling="2025-10-01 09:49:17.705980566 +0000 UTC m=+789.821124723" observedRunningTime="2025-10-01 09:49:18.055440248 +0000 UTC m=+790.170584405" watchObservedRunningTime="2025-10-01 09:49:18.059706048 +0000 UTC m=+790.174850205" Oct 01 09:49:18 crc kubenswrapper[4787]: I1001 09:49:18.611196 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5tmfr"] Oct 01 09:49:19 crc kubenswrapper[4787]: I1001 09:49:19.033588 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5tmfr" podUID="c95df987-c9ac-4bc2-9077-30cb86f3fa7b" containerName="registry-server" containerID="cri-o://ce11f464aa1b3fe71d81c2de46eb18219d477d4cc54163d445e2734353941a35" gracePeriod=2 Oct 01 09:49:19 crc kubenswrapper[4787]: I1001 09:49:19.536158 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5tmfr" Oct 01 09:49:19 crc kubenswrapper[4787]: I1001 09:49:19.622177 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c95df987-c9ac-4bc2-9077-30cb86f3fa7b-utilities\") pod \"c95df987-c9ac-4bc2-9077-30cb86f3fa7b\" (UID: \"c95df987-c9ac-4bc2-9077-30cb86f3fa7b\") " Oct 01 09:49:19 crc kubenswrapper[4787]: I1001 09:49:19.622229 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c95df987-c9ac-4bc2-9077-30cb86f3fa7b-catalog-content\") pod \"c95df987-c9ac-4bc2-9077-30cb86f3fa7b\" (UID: \"c95df987-c9ac-4bc2-9077-30cb86f3fa7b\") " Oct 01 09:49:19 crc kubenswrapper[4787]: I1001 09:49:19.622255 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44tmb\" (UniqueName: \"kubernetes.io/projected/c95df987-c9ac-4bc2-9077-30cb86f3fa7b-kube-api-access-44tmb\") pod \"c95df987-c9ac-4bc2-9077-30cb86f3fa7b\" (UID: \"c95df987-c9ac-4bc2-9077-30cb86f3fa7b\") " Oct 01 09:49:19 crc kubenswrapper[4787]: I1001 09:49:19.623112 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c95df987-c9ac-4bc2-9077-30cb86f3fa7b-utilities" (OuterVolumeSpecName: "utilities") pod "c95df987-c9ac-4bc2-9077-30cb86f3fa7b" (UID: "c95df987-c9ac-4bc2-9077-30cb86f3fa7b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:49:19 crc kubenswrapper[4787]: I1001 09:49:19.629220 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c95df987-c9ac-4bc2-9077-30cb86f3fa7b-kube-api-access-44tmb" (OuterVolumeSpecName: "kube-api-access-44tmb") pod "c95df987-c9ac-4bc2-9077-30cb86f3fa7b" (UID: "c95df987-c9ac-4bc2-9077-30cb86f3fa7b"). InnerVolumeSpecName "kube-api-access-44tmb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:49:19 crc kubenswrapper[4787]: I1001 09:49:19.704304 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c95df987-c9ac-4bc2-9077-30cb86f3fa7b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c95df987-c9ac-4bc2-9077-30cb86f3fa7b" (UID: "c95df987-c9ac-4bc2-9077-30cb86f3fa7b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:49:19 crc kubenswrapper[4787]: I1001 09:49:19.723829 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c95df987-c9ac-4bc2-9077-30cb86f3fa7b-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:49:19 crc kubenswrapper[4787]: I1001 09:49:19.723865 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c95df987-c9ac-4bc2-9077-30cb86f3fa7b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:49:19 crc kubenswrapper[4787]: I1001 09:49:19.723876 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44tmb\" (UniqueName: \"kubernetes.io/projected/c95df987-c9ac-4bc2-9077-30cb86f3fa7b-kube-api-access-44tmb\") on node \"crc\" DevicePath \"\"" Oct 01 09:49:20 crc kubenswrapper[4787]: I1001 09:49:20.041709 4787 generic.go:334] "Generic (PLEG): container finished" podID="c95df987-c9ac-4bc2-9077-30cb86f3fa7b" containerID="ce11f464aa1b3fe71d81c2de46eb18219d477d4cc54163d445e2734353941a35" exitCode=0 Oct 01 09:49:20 crc kubenswrapper[4787]: I1001 09:49:20.041752 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5tmfr" event={"ID":"c95df987-c9ac-4bc2-9077-30cb86f3fa7b","Type":"ContainerDied","Data":"ce11f464aa1b3fe71d81c2de46eb18219d477d4cc54163d445e2734353941a35"} Oct 01 09:49:20 crc kubenswrapper[4787]: I1001 09:49:20.041782 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5tmfr" event={"ID":"c95df987-c9ac-4bc2-9077-30cb86f3fa7b","Type":"ContainerDied","Data":"d983ca32a5a6e0bef7a14942bc2d85eaeec06a586df6821edafc001c5c4968a0"} Oct 01 09:49:20 crc kubenswrapper[4787]: I1001 09:49:20.041802 4787 scope.go:117] "RemoveContainer" containerID="ce11f464aa1b3fe71d81c2de46eb18219d477d4cc54163d445e2734353941a35" Oct 01 09:49:20 crc kubenswrapper[4787]: I1001 09:49:20.041877 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5tmfr" Oct 01 09:49:20 crc kubenswrapper[4787]: I1001 09:49:20.070870 4787 scope.go:117] "RemoveContainer" containerID="9bbfbb7da6ab1c8609b2283243e1cb185a445a698d7cace7a4b479db6809fc15" Oct 01 09:49:20 crc kubenswrapper[4787]: I1001 09:49:20.095554 4787 scope.go:117] "RemoveContainer" containerID="3f46cd8dcfae5b66f12dd446fd8722d0470a69575d836308c0c93f59bc392ec6" Oct 01 09:49:20 crc kubenswrapper[4787]: I1001 09:49:20.139093 4787 scope.go:117] "RemoveContainer" containerID="ce11f464aa1b3fe71d81c2de46eb18219d477d4cc54163d445e2734353941a35" Oct 01 09:49:20 crc kubenswrapper[4787]: E1001 09:49:20.139722 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce11f464aa1b3fe71d81c2de46eb18219d477d4cc54163d445e2734353941a35\": container with ID starting with ce11f464aa1b3fe71d81c2de46eb18219d477d4cc54163d445e2734353941a35 not found: ID does not exist" containerID="ce11f464aa1b3fe71d81c2de46eb18219d477d4cc54163d445e2734353941a35" Oct 01 09:49:20 crc kubenswrapper[4787]: I1001 09:49:20.139795 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce11f464aa1b3fe71d81c2de46eb18219d477d4cc54163d445e2734353941a35"} err="failed to get container status \"ce11f464aa1b3fe71d81c2de46eb18219d477d4cc54163d445e2734353941a35\": rpc error: code = NotFound desc = could not find container \"ce11f464aa1b3fe71d81c2de46eb18219d477d4cc54163d445e2734353941a35\": container with ID starting with ce11f464aa1b3fe71d81c2de46eb18219d477d4cc54163d445e2734353941a35 not found: ID does not exist" Oct 01 09:49:20 crc kubenswrapper[4787]: I1001 09:49:20.139867 4787 scope.go:117] "RemoveContainer" containerID="9bbfbb7da6ab1c8609b2283243e1cb185a445a698d7cace7a4b479db6809fc15" Oct 01 09:49:20 crc kubenswrapper[4787]: I1001 09:49:20.140653 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5tmfr"] Oct 01 09:49:20 crc kubenswrapper[4787]: E1001 09:49:20.140847 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bbfbb7da6ab1c8609b2283243e1cb185a445a698d7cace7a4b479db6809fc15\": container with ID starting with 9bbfbb7da6ab1c8609b2283243e1cb185a445a698d7cace7a4b479db6809fc15 not found: ID does not exist" containerID="9bbfbb7da6ab1c8609b2283243e1cb185a445a698d7cace7a4b479db6809fc15" Oct 01 09:49:20 crc kubenswrapper[4787]: I1001 09:49:20.141017 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bbfbb7da6ab1c8609b2283243e1cb185a445a698d7cace7a4b479db6809fc15"} err="failed to get container status \"9bbfbb7da6ab1c8609b2283243e1cb185a445a698d7cace7a4b479db6809fc15\": rpc error: code = NotFound desc = could not find container \"9bbfbb7da6ab1c8609b2283243e1cb185a445a698d7cace7a4b479db6809fc15\": container with ID starting with 9bbfbb7da6ab1c8609b2283243e1cb185a445a698d7cace7a4b479db6809fc15 not found: ID does not exist" Oct 01 09:49:20 crc kubenswrapper[4787]: I1001 09:49:20.141397 4787 scope.go:117] "RemoveContainer" containerID="3f46cd8dcfae5b66f12dd446fd8722d0470a69575d836308c0c93f59bc392ec6" Oct 01 09:49:20 crc kubenswrapper[4787]: E1001 09:49:20.142410 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f46cd8dcfae5b66f12dd446fd8722d0470a69575d836308c0c93f59bc392ec6\": container with ID starting with 3f46cd8dcfae5b66f12dd446fd8722d0470a69575d836308c0c93f59bc392ec6 not found: ID does not exist" containerID="3f46cd8dcfae5b66f12dd446fd8722d0470a69575d836308c0c93f59bc392ec6" Oct 01 09:49:20 crc kubenswrapper[4787]: I1001 09:49:20.142463 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f46cd8dcfae5b66f12dd446fd8722d0470a69575d836308c0c93f59bc392ec6"} err="failed to get container status \"3f46cd8dcfae5b66f12dd446fd8722d0470a69575d836308c0c93f59bc392ec6\": rpc error: code = NotFound desc = could not find container \"3f46cd8dcfae5b66f12dd446fd8722d0470a69575d836308c0c93f59bc392ec6\": container with ID starting with 3f46cd8dcfae5b66f12dd446fd8722d0470a69575d836308c0c93f59bc392ec6 not found: ID does not exist" Oct 01 09:49:20 crc kubenswrapper[4787]: I1001 09:49:20.144771 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5tmfr"] Oct 01 09:49:20 crc kubenswrapper[4787]: I1001 09:49:20.533751 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c95df987-c9ac-4bc2-9077-30cb86f3fa7b" path="/var/lib/kubelet/pods/c95df987-c9ac-4bc2-9077-30cb86f3fa7b/volumes" Oct 01 09:49:20 crc kubenswrapper[4787]: I1001 09:49:20.672777 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-676c66f88b-7hgdc" Oct 01 09:49:32 crc kubenswrapper[4787]: I1001 09:49:32.223222 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-828zk"] Oct 01 09:49:32 crc kubenswrapper[4787]: E1001 09:49:32.224118 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c95df987-c9ac-4bc2-9077-30cb86f3fa7b" containerName="extract-content" Oct 01 09:49:32 crc kubenswrapper[4787]: I1001 09:49:32.224136 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c95df987-c9ac-4bc2-9077-30cb86f3fa7b" containerName="extract-content" Oct 01 09:49:32 crc kubenswrapper[4787]: E1001 09:49:32.224164 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c95df987-c9ac-4bc2-9077-30cb86f3fa7b" containerName="extract-utilities" Oct 01 09:49:32 crc kubenswrapper[4787]: I1001 09:49:32.224173 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c95df987-c9ac-4bc2-9077-30cb86f3fa7b" containerName="extract-utilities" Oct 01 09:49:32 crc kubenswrapper[4787]: E1001 09:49:32.224185 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c95df987-c9ac-4bc2-9077-30cb86f3fa7b" containerName="registry-server" Oct 01 09:49:32 crc kubenswrapper[4787]: I1001 09:49:32.224193 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c95df987-c9ac-4bc2-9077-30cb86f3fa7b" containerName="registry-server" Oct 01 09:49:32 crc kubenswrapper[4787]: I1001 09:49:32.224320 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="c95df987-c9ac-4bc2-9077-30cb86f3fa7b" containerName="registry-server" Oct 01 09:49:32 crc kubenswrapper[4787]: I1001 09:49:32.225122 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-828zk" Oct 01 09:49:32 crc kubenswrapper[4787]: I1001 09:49:32.236204 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-828zk"] Oct 01 09:49:32 crc kubenswrapper[4787]: I1001 09:49:32.296544 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce71de87-f660-4d9e-845c-1f722a75aa54-catalog-content\") pod \"redhat-marketplace-828zk\" (UID: \"ce71de87-f660-4d9e-845c-1f722a75aa54\") " pod="openshift-marketplace/redhat-marketplace-828zk" Oct 01 09:49:32 crc kubenswrapper[4787]: I1001 09:49:32.296607 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhm7t\" (UniqueName: \"kubernetes.io/projected/ce71de87-f660-4d9e-845c-1f722a75aa54-kube-api-access-lhm7t\") pod \"redhat-marketplace-828zk\" (UID: \"ce71de87-f660-4d9e-845c-1f722a75aa54\") " pod="openshift-marketplace/redhat-marketplace-828zk" Oct 01 09:49:32 crc kubenswrapper[4787]: I1001 09:49:32.296630 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce71de87-f660-4d9e-845c-1f722a75aa54-utilities\") pod \"redhat-marketplace-828zk\" (UID: \"ce71de87-f660-4d9e-845c-1f722a75aa54\") " pod="openshift-marketplace/redhat-marketplace-828zk" Oct 01 09:49:32 crc kubenswrapper[4787]: I1001 09:49:32.398381 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce71de87-f660-4d9e-845c-1f722a75aa54-catalog-content\") pod \"redhat-marketplace-828zk\" (UID: \"ce71de87-f660-4d9e-845c-1f722a75aa54\") " pod="openshift-marketplace/redhat-marketplace-828zk" Oct 01 09:49:32 crc kubenswrapper[4787]: I1001 09:49:32.398432 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhm7t\" (UniqueName: \"kubernetes.io/projected/ce71de87-f660-4d9e-845c-1f722a75aa54-kube-api-access-lhm7t\") pod \"redhat-marketplace-828zk\" (UID: \"ce71de87-f660-4d9e-845c-1f722a75aa54\") " pod="openshift-marketplace/redhat-marketplace-828zk" Oct 01 09:49:32 crc kubenswrapper[4787]: I1001 09:49:32.398457 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce71de87-f660-4d9e-845c-1f722a75aa54-utilities\") pod \"redhat-marketplace-828zk\" (UID: \"ce71de87-f660-4d9e-845c-1f722a75aa54\") " pod="openshift-marketplace/redhat-marketplace-828zk" Oct 01 09:49:32 crc kubenswrapper[4787]: I1001 09:49:32.398992 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce71de87-f660-4d9e-845c-1f722a75aa54-utilities\") pod \"redhat-marketplace-828zk\" (UID: \"ce71de87-f660-4d9e-845c-1f722a75aa54\") " pod="openshift-marketplace/redhat-marketplace-828zk" Oct 01 09:49:32 crc kubenswrapper[4787]: I1001 09:49:32.399416 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce71de87-f660-4d9e-845c-1f722a75aa54-catalog-content\") pod \"redhat-marketplace-828zk\" (UID: \"ce71de87-f660-4d9e-845c-1f722a75aa54\") " pod="openshift-marketplace/redhat-marketplace-828zk" Oct 01 09:49:32 crc kubenswrapper[4787]: I1001 09:49:32.434507 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhm7t\" (UniqueName: \"kubernetes.io/projected/ce71de87-f660-4d9e-845c-1f722a75aa54-kube-api-access-lhm7t\") pod \"redhat-marketplace-828zk\" (UID: \"ce71de87-f660-4d9e-845c-1f722a75aa54\") " pod="openshift-marketplace/redhat-marketplace-828zk" Oct 01 09:49:32 crc kubenswrapper[4787]: I1001 09:49:32.541948 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-828zk" Oct 01 09:49:33 crc kubenswrapper[4787]: I1001 09:49:33.038446 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-828zk"] Oct 01 09:49:33 crc kubenswrapper[4787]: W1001 09:49:33.048674 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podce71de87_f660_4d9e_845c_1f722a75aa54.slice/crio-eb6899c1145e1df26da8bf8722cbbe6fbfc598e099e31c07ede2d96eb8e80bc3 WatchSource:0}: Error finding container eb6899c1145e1df26da8bf8722cbbe6fbfc598e099e31c07ede2d96eb8e80bc3: Status 404 returned error can't find the container with id eb6899c1145e1df26da8bf8722cbbe6fbfc598e099e31c07ede2d96eb8e80bc3 Oct 01 09:49:33 crc kubenswrapper[4787]: I1001 09:49:33.129925 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-828zk" event={"ID":"ce71de87-f660-4d9e-845c-1f722a75aa54","Type":"ContainerStarted","Data":"eb6899c1145e1df26da8bf8722cbbe6fbfc598e099e31c07ede2d96eb8e80bc3"} Oct 01 09:49:34 crc kubenswrapper[4787]: I1001 09:49:34.143110 4787 generic.go:334] "Generic (PLEG): container finished" podID="ce71de87-f660-4d9e-845c-1f722a75aa54" containerID="7104fc1125c4f6296ec1f7983a523d7e97b7bc3be97f62428055208974c5d9f4" exitCode=0 Oct 01 09:49:34 crc kubenswrapper[4787]: I1001 09:49:34.143542 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-828zk" event={"ID":"ce71de87-f660-4d9e-845c-1f722a75aa54","Type":"ContainerDied","Data":"7104fc1125c4f6296ec1f7983a523d7e97b7bc3be97f62428055208974c5d9f4"} Oct 01 09:49:35 crc kubenswrapper[4787]: I1001 09:49:35.152889 4787 generic.go:334] "Generic (PLEG): container finished" podID="ce71de87-f660-4d9e-845c-1f722a75aa54" containerID="283a38eebc39e51af29e296289ac06e4f14a43d346a8fa60307b3521e2d6eab2" exitCode=0 Oct 01 09:49:35 crc kubenswrapper[4787]: I1001 09:49:35.152962 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-828zk" event={"ID":"ce71de87-f660-4d9e-845c-1f722a75aa54","Type":"ContainerDied","Data":"283a38eebc39e51af29e296289ac06e4f14a43d346a8fa60307b3521e2d6eab2"} Oct 01 09:49:36 crc kubenswrapper[4787]: I1001 09:49:36.164160 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-828zk" event={"ID":"ce71de87-f660-4d9e-845c-1f722a75aa54","Type":"ContainerStarted","Data":"690a760a08ddfc9c719b917a5782b9472b36455248abaf828731f30bb471b8e4"} Oct 01 09:49:36 crc kubenswrapper[4787]: I1001 09:49:36.190983 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-828zk" podStartSLOduration=2.768586671 podStartE2EDuration="4.190958097s" podCreationTimestamp="2025-10-01 09:49:32 +0000 UTC" firstStartedPulling="2025-10-01 09:49:34.145736975 +0000 UTC m=+806.260881132" lastFinishedPulling="2025-10-01 09:49:35.568108401 +0000 UTC m=+807.683252558" observedRunningTime="2025-10-01 09:49:36.186372349 +0000 UTC m=+808.301516526" watchObservedRunningTime="2025-10-01 09:49:36.190958097 +0000 UTC m=+808.306102254" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.018519 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859cd486d-6h8cm"] Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.020695 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859cd486d-6h8cm" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.023340 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-f7f98cb69-wqs6n"] Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.025362 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-wqs6n" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.026899 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-9jr6l" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.033455 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-94f48" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.054015 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859cd486d-6h8cm"] Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.087727 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-f7f98cb69-wqs6n"] Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.091420 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mk2p\" (UniqueName: \"kubernetes.io/projected/26b292ae-dcb2-41f6-86b5-2eddfecdf356-kube-api-access-4mk2p\") pod \"barbican-operator-controller-manager-f7f98cb69-wqs6n\" (UID: \"26b292ae-dcb2-41f6-86b5-2eddfecdf356\") " pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-wqs6n" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.091524 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bw6n\" (UniqueName: \"kubernetes.io/projected/24354b04-0daa-4fd4-8d38-b21f64bc2cfe-kube-api-access-5bw6n\") pod \"cinder-operator-controller-manager-859cd486d-6h8cm\" (UID: \"24354b04-0daa-4fd4-8d38-b21f64bc2cfe\") " pod="openstack-operators/cinder-operator-controller-manager-859cd486d-6h8cm" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.102814 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-77fb7bcf5b-2pxzs"] Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.104093 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-2pxzs" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.112154 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-6rdlx" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.132636 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-77fb7bcf5b-2pxzs"] Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.143910 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-8bc4775b5-x2v89"] Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.145293 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-x2v89" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.150349 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5b4fc86755-lbbxd"] Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.151866 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-lbbxd" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.157631 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-njtqw" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.157923 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-hmc4v" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.161461 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-8bc4775b5-x2v89"] Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.176235 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-679b4759bb-qt9bq"] Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.177604 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-qt9bq" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.179707 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-t72zp"] Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.181303 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-t72zp" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.194428 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-k72s7" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.194657 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.195114 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-hp86k" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.196166 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bw6n\" (UniqueName: \"kubernetes.io/projected/24354b04-0daa-4fd4-8d38-b21f64bc2cfe-kube-api-access-5bw6n\") pod \"cinder-operator-controller-manager-859cd486d-6h8cm\" (UID: \"24354b04-0daa-4fd4-8d38-b21f64bc2cfe\") " pod="openstack-operators/cinder-operator-controller-manager-859cd486d-6h8cm" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.196235 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dr74\" (UniqueName: \"kubernetes.io/projected/8f52292a-2603-4524-b3fc-8d7b003f95f8-kube-api-access-4dr74\") pod \"horizon-operator-controller-manager-679b4759bb-qt9bq\" (UID: \"8f52292a-2603-4524-b3fc-8d7b003f95f8\") " pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-qt9bq" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.196264 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xvgw\" (UniqueName: \"kubernetes.io/projected/9e03d54f-d462-4894-9c3b-8253a861af54-kube-api-access-6xvgw\") pod \"infra-operator-controller-manager-5c8fdc4d5c-t72zp\" (UID: \"9e03d54f-d462-4894-9c3b-8253a861af54\") " pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-t72zp" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.196288 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9e03d54f-d462-4894-9c3b-8253a861af54-cert\") pod \"infra-operator-controller-manager-5c8fdc4d5c-t72zp\" (UID: \"9e03d54f-d462-4894-9c3b-8253a861af54\") " pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-t72zp" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.196343 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mk2p\" (UniqueName: \"kubernetes.io/projected/26b292ae-dcb2-41f6-86b5-2eddfecdf356-kube-api-access-4mk2p\") pod \"barbican-operator-controller-manager-f7f98cb69-wqs6n\" (UID: \"26b292ae-dcb2-41f6-86b5-2eddfecdf356\") " pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-wqs6n" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.196373 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nlnr\" (UniqueName: \"kubernetes.io/projected/b1a0ae18-3c6d-473a-a660-5ff59f6a87bf-kube-api-access-4nlnr\") pod \"heat-operator-controller-manager-5b4fc86755-lbbxd\" (UID: \"b1a0ae18-3c6d-473a-a660-5ff59f6a87bf\") " pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-lbbxd" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.196398 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7r6z\" (UniqueName: \"kubernetes.io/projected/77e07ae2-1dae-4570-84b8-c137e86ac628-kube-api-access-q7r6z\") pod \"designate-operator-controller-manager-77fb7bcf5b-2pxzs\" (UID: \"77e07ae2-1dae-4570-84b8-c137e86ac628\") " pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-2pxzs" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.196428 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9t58q\" (UniqueName: \"kubernetes.io/projected/b4902328-fb90-40f9-a84c-5a0359d699d4-kube-api-access-9t58q\") pod \"glance-operator-controller-manager-8bc4775b5-x2v89\" (UID: \"b4902328-fb90-40f9-a84c-5a0359d699d4\") " pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-x2v89" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.202066 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5b4fc86755-lbbxd"] Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.221249 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-679b4759bb-qt9bq"] Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.231650 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bw6n\" (UniqueName: \"kubernetes.io/projected/24354b04-0daa-4fd4-8d38-b21f64bc2cfe-kube-api-access-5bw6n\") pod \"cinder-operator-controller-manager-859cd486d-6h8cm\" (UID: \"24354b04-0daa-4fd4-8d38-b21f64bc2cfe\") " pod="openstack-operators/cinder-operator-controller-manager-859cd486d-6h8cm" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.245765 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mk2p\" (UniqueName: \"kubernetes.io/projected/26b292ae-dcb2-41f6-86b5-2eddfecdf356-kube-api-access-4mk2p\") pod \"barbican-operator-controller-manager-f7f98cb69-wqs6n\" (UID: \"26b292ae-dcb2-41f6-86b5-2eddfecdf356\") " pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-wqs6n" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.253666 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-t72zp"] Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.264580 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5f45cd594f-zrtv2"] Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.265548 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5f45cd594f-zrtv2" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.280960 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-z56qk" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.281117 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5f45cd594f-zrtv2"] Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.297636 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dr74\" (UniqueName: \"kubernetes.io/projected/8f52292a-2603-4524-b3fc-8d7b003f95f8-kube-api-access-4dr74\") pod \"horizon-operator-controller-manager-679b4759bb-qt9bq\" (UID: \"8f52292a-2603-4524-b3fc-8d7b003f95f8\") " pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-qt9bq" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.297673 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xvgw\" (UniqueName: \"kubernetes.io/projected/9e03d54f-d462-4894-9c3b-8253a861af54-kube-api-access-6xvgw\") pod \"infra-operator-controller-manager-5c8fdc4d5c-t72zp\" (UID: \"9e03d54f-d462-4894-9c3b-8253a861af54\") " pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-t72zp" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.297695 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9e03d54f-d462-4894-9c3b-8253a861af54-cert\") pod \"infra-operator-controller-manager-5c8fdc4d5c-t72zp\" (UID: \"9e03d54f-d462-4894-9c3b-8253a861af54\") " pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-t72zp" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.297743 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nlnr\" (UniqueName: \"kubernetes.io/projected/b1a0ae18-3c6d-473a-a660-5ff59f6a87bf-kube-api-access-4nlnr\") pod \"heat-operator-controller-manager-5b4fc86755-lbbxd\" (UID: \"b1a0ae18-3c6d-473a-a660-5ff59f6a87bf\") " pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-lbbxd" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.297769 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7r6z\" (UniqueName: \"kubernetes.io/projected/77e07ae2-1dae-4570-84b8-c137e86ac628-kube-api-access-q7r6z\") pod \"designate-operator-controller-manager-77fb7bcf5b-2pxzs\" (UID: \"77e07ae2-1dae-4570-84b8-c137e86ac628\") " pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-2pxzs" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.297820 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9t58q\" (UniqueName: \"kubernetes.io/projected/b4902328-fb90-40f9-a84c-5a0359d699d4-kube-api-access-9t58q\") pod \"glance-operator-controller-manager-8bc4775b5-x2v89\" (UID: \"b4902328-fb90-40f9-a84c-5a0359d699d4\") " pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-x2v89" Oct 01 09:49:37 crc kubenswrapper[4787]: E1001 09:49:37.300552 4787 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 01 09:49:37 crc kubenswrapper[4787]: E1001 09:49:37.300610 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9e03d54f-d462-4894-9c3b-8253a861af54-cert podName:9e03d54f-d462-4894-9c3b-8253a861af54 nodeName:}" failed. No retries permitted until 2025-10-01 09:49:37.800592238 +0000 UTC m=+809.915736395 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/9e03d54f-d462-4894-9c3b-8253a861af54-cert") pod "infra-operator-controller-manager-5c8fdc4d5c-t72zp" (UID: "9e03d54f-d462-4894-9c3b-8253a861af54") : secret "infra-operator-webhook-server-cert" not found Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.317845 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-59d7dc95cf-9tdp2"] Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.318835 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-9tdp2" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.331586 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-klllr" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.336293 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xvgw\" (UniqueName: \"kubernetes.io/projected/9e03d54f-d462-4894-9c3b-8253a861af54-kube-api-access-6xvgw\") pod \"infra-operator-controller-manager-5c8fdc4d5c-t72zp\" (UID: \"9e03d54f-d462-4894-9c3b-8253a861af54\") " pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-t72zp" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.336434 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nlnr\" (UniqueName: \"kubernetes.io/projected/b1a0ae18-3c6d-473a-a660-5ff59f6a87bf-kube-api-access-4nlnr\") pod \"heat-operator-controller-manager-5b4fc86755-lbbxd\" (UID: \"b1a0ae18-3c6d-473a-a660-5ff59f6a87bf\") " pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-lbbxd" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.349234 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-b7cf8cb5f-5qz2m"] Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.349738 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7r6z\" (UniqueName: \"kubernetes.io/projected/77e07ae2-1dae-4570-84b8-c137e86ac628-kube-api-access-q7r6z\") pod \"designate-operator-controller-manager-77fb7bcf5b-2pxzs\" (UID: \"77e07ae2-1dae-4570-84b8-c137e86ac628\") " pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-2pxzs" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.350319 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-5qz2m" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.355199 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-gq866" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.355877 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-59d7dc95cf-9tdp2"] Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.358526 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859cd486d-6h8cm" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.364641 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dr74\" (UniqueName: \"kubernetes.io/projected/8f52292a-2603-4524-b3fc-8d7b003f95f8-kube-api-access-4dr74\") pod \"horizon-operator-controller-manager-679b4759bb-qt9bq\" (UID: \"8f52292a-2603-4524-b3fc-8d7b003f95f8\") " pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-qt9bq" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.380250 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-b7cf8cb5f-5qz2m"] Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.380624 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9t58q\" (UniqueName: \"kubernetes.io/projected/b4902328-fb90-40f9-a84c-5a0359d699d4-kube-api-access-9t58q\") pod \"glance-operator-controller-manager-8bc4775b5-x2v89\" (UID: \"b4902328-fb90-40f9-a84c-5a0359d699d4\") " pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-x2v89" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.384480 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-wqs6n" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.392755 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67bf5bb885-qmr4v"] Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.393904 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-qmr4v" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.399913 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4mvm\" (UniqueName: \"kubernetes.io/projected/ae2ba2f0-8422-4a03-b64f-d335f76c5532-kube-api-access-n4mvm\") pod \"ironic-operator-controller-manager-5f45cd594f-zrtv2\" (UID: \"ae2ba2f0-8422-4a03-b64f-d335f76c5532\") " pod="openstack-operators/ironic-operator-controller-manager-5f45cd594f-zrtv2" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.401241 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-nbmx4" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.438581 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-2pxzs" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.449258 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-54fbbfcd44-4jlcz"] Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.450664 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-54fbbfcd44-4jlcz" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.473996 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-g9vwr" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.475107 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-x2v89" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.485848 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67bf5bb885-qmr4v"] Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.502590 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4mvm\" (UniqueName: \"kubernetes.io/projected/ae2ba2f0-8422-4a03-b64f-d335f76c5532-kube-api-access-n4mvm\") pod \"ironic-operator-controller-manager-5f45cd594f-zrtv2\" (UID: \"ae2ba2f0-8422-4a03-b64f-d335f76c5532\") " pod="openstack-operators/ironic-operator-controller-manager-5f45cd594f-zrtv2" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.502958 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tdhg\" (UniqueName: \"kubernetes.io/projected/88d93cbc-27c2-47be-a882-6607925d4391-kube-api-access-6tdhg\") pod \"keystone-operator-controller-manager-59d7dc95cf-9tdp2\" (UID: \"88d93cbc-27c2-47be-a882-6607925d4391\") " pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-9tdp2" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.503088 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmc24\" (UniqueName: \"kubernetes.io/projected/8b59a5e1-4f45-4bcd-b9ce-63bae8b5ed57-kube-api-access-rmc24\") pod \"neutron-operator-controller-manager-54fbbfcd44-4jlcz\" (UID: \"8b59a5e1-4f45-4bcd-b9ce-63bae8b5ed57\") " pod="openstack-operators/neutron-operator-controller-manager-54fbbfcd44-4jlcz" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.503221 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xbq4\" (UniqueName: \"kubernetes.io/projected/2dfc2636-198f-4b51-b07e-40ff76c1091c-kube-api-access-8xbq4\") pod \"manila-operator-controller-manager-b7cf8cb5f-5qz2m\" (UID: \"2dfc2636-198f-4b51-b07e-40ff76c1091c\") " pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-5qz2m" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.503316 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-442mm\" (UniqueName: \"kubernetes.io/projected/b54a691a-cd96-4ae4-a9ce-65662f58cd60-kube-api-access-442mm\") pod \"mariadb-operator-controller-manager-67bf5bb885-qmr4v\" (UID: \"b54a691a-cd96-4ae4-a9ce-65662f58cd60\") " pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-qmr4v" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.512824 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-54fbbfcd44-4jlcz"] Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.513627 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-lbbxd" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.527413 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-qt9bq" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.551791 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4mvm\" (UniqueName: \"kubernetes.io/projected/ae2ba2f0-8422-4a03-b64f-d335f76c5532-kube-api-access-n4mvm\") pod \"ironic-operator-controller-manager-5f45cd594f-zrtv2\" (UID: \"ae2ba2f0-8422-4a03-b64f-d335f76c5532\") " pod="openstack-operators/ironic-operator-controller-manager-5f45cd594f-zrtv2" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.586143 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-nbxwx"] Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.587329 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-nbxwx" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.591205 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-nbxwx"] Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.595158 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-75f8d67d86-xjbwt"] Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.596699 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-75f8d67d86-xjbwt" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.597498 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-t8tpm" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.611818 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-75f8d67d86-xjbwt"] Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.625447 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9phq\" (UniqueName: \"kubernetes.io/projected/c186a6f1-61b9-4ab7-8996-5e1d8439a77c-kube-api-access-m9phq\") pod \"nova-operator-controller-manager-7fd5b6bbc6-nbxwx\" (UID: \"c186a6f1-61b9-4ab7-8996-5e1d8439a77c\") " pod="openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-nbxwx" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.628577 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5f45cd594f-zrtv2" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.629414 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xbq4\" (UniqueName: \"kubernetes.io/projected/2dfc2636-198f-4b51-b07e-40ff76c1091c-kube-api-access-8xbq4\") pod \"manila-operator-controller-manager-b7cf8cb5f-5qz2m\" (UID: \"2dfc2636-198f-4b51-b07e-40ff76c1091c\") " pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-5qz2m" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.629463 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-442mm\" (UniqueName: \"kubernetes.io/projected/b54a691a-cd96-4ae4-a9ce-65662f58cd60-kube-api-access-442mm\") pod \"mariadb-operator-controller-manager-67bf5bb885-qmr4v\" (UID: \"b54a691a-cd96-4ae4-a9ce-65662f58cd60\") " pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-qmr4v" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.629516 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79hq8\" (UniqueName: \"kubernetes.io/projected/20e0b9a6-f51b-4925-bffd-8fa00e9a8ec7-kube-api-access-79hq8\") pod \"octavia-operator-controller-manager-75f8d67d86-xjbwt\" (UID: \"20e0b9a6-f51b-4925-bffd-8fa00e9a8ec7\") " pod="openstack-operators/octavia-operator-controller-manager-75f8d67d86-xjbwt" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.629578 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tdhg\" (UniqueName: \"kubernetes.io/projected/88d93cbc-27c2-47be-a882-6607925d4391-kube-api-access-6tdhg\") pod \"keystone-operator-controller-manager-59d7dc95cf-9tdp2\" (UID: \"88d93cbc-27c2-47be-a882-6607925d4391\") " pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-9tdp2" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.629600 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmc24\" (UniqueName: \"kubernetes.io/projected/8b59a5e1-4f45-4bcd-b9ce-63bae8b5ed57-kube-api-access-rmc24\") pod \"neutron-operator-controller-manager-54fbbfcd44-4jlcz\" (UID: \"8b59a5e1-4f45-4bcd-b9ce-63bae8b5ed57\") " pod="openstack-operators/neutron-operator-controller-manager-54fbbfcd44-4jlcz" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.635817 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-42jqf" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.702007 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-659bb84579xswrt"] Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.711346 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmc24\" (UniqueName: \"kubernetes.io/projected/8b59a5e1-4f45-4bcd-b9ce-63bae8b5ed57-kube-api-access-rmc24\") pod \"neutron-operator-controller-manager-54fbbfcd44-4jlcz\" (UID: \"8b59a5e1-4f45-4bcd-b9ce-63bae8b5ed57\") " pod="openstack-operators/neutron-operator-controller-manager-54fbbfcd44-4jlcz" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.713023 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-659bb84579xswrt" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.725874 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-54ktq" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.731794 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-684t7\" (UniqueName: \"kubernetes.io/projected/2d5ffd79-67fa-4dd0-81a7-b44156946b86-kube-api-access-684t7\") pod \"openstack-baremetal-operator-controller-manager-659bb84579xswrt\" (UID: \"2d5ffd79-67fa-4dd0-81a7-b44156946b86\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-659bb84579xswrt" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.731899 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9phq\" (UniqueName: \"kubernetes.io/projected/c186a6f1-61b9-4ab7-8996-5e1d8439a77c-kube-api-access-m9phq\") pod \"nova-operator-controller-manager-7fd5b6bbc6-nbxwx\" (UID: \"c186a6f1-61b9-4ab7-8996-5e1d8439a77c\") " pod="openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-nbxwx" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.731945 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2d5ffd79-67fa-4dd0-81a7-b44156946b86-cert\") pod \"openstack-baremetal-operator-controller-manager-659bb84579xswrt\" (UID: \"2d5ffd79-67fa-4dd0-81a7-b44156946b86\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-659bb84579xswrt" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.732025 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79hq8\" (UniqueName: \"kubernetes.io/projected/20e0b9a6-f51b-4925-bffd-8fa00e9a8ec7-kube-api-access-79hq8\") pod \"octavia-operator-controller-manager-75f8d67d86-xjbwt\" (UID: \"20e0b9a6-f51b-4925-bffd-8fa00e9a8ec7\") " pod="openstack-operators/octavia-operator-controller-manager-75f8d67d86-xjbwt" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.736505 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-442mm\" (UniqueName: \"kubernetes.io/projected/b54a691a-cd96-4ae4-a9ce-65662f58cd60-kube-api-access-442mm\") pod \"mariadb-operator-controller-manager-67bf5bb885-qmr4v\" (UID: \"b54a691a-cd96-4ae4-a9ce-65662f58cd60\") " pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-qmr4v" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.786811 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xbq4\" (UniqueName: \"kubernetes.io/projected/2dfc2636-198f-4b51-b07e-40ff76c1091c-kube-api-access-8xbq4\") pod \"manila-operator-controller-manager-b7cf8cb5f-5qz2m\" (UID: \"2dfc2636-198f-4b51-b07e-40ff76c1091c\") " pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-5qz2m" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.795410 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.811595 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-qmr4v" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.811987 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-54fbbfcd44-4jlcz" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.849340 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79hq8\" (UniqueName: \"kubernetes.io/projected/20e0b9a6-f51b-4925-bffd-8fa00e9a8ec7-kube-api-access-79hq8\") pod \"octavia-operator-controller-manager-75f8d67d86-xjbwt\" (UID: \"20e0b9a6-f51b-4925-bffd-8fa00e9a8ec7\") " pod="openstack-operators/octavia-operator-controller-manager-75f8d67d86-xjbwt" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.891507 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tdhg\" (UniqueName: \"kubernetes.io/projected/88d93cbc-27c2-47be-a882-6607925d4391-kube-api-access-6tdhg\") pod \"keystone-operator-controller-manager-59d7dc95cf-9tdp2\" (UID: \"88d93cbc-27c2-47be-a882-6607925d4391\") " pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-9tdp2" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.891961 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9phq\" (UniqueName: \"kubernetes.io/projected/c186a6f1-61b9-4ab7-8996-5e1d8439a77c-kube-api-access-m9phq\") pod \"nova-operator-controller-manager-7fd5b6bbc6-nbxwx\" (UID: \"c186a6f1-61b9-4ab7-8996-5e1d8439a77c\") " pod="openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-nbxwx" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.893290 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9e03d54f-d462-4894-9c3b-8253a861af54-cert\") pod \"infra-operator-controller-manager-5c8fdc4d5c-t72zp\" (UID: \"9e03d54f-d462-4894-9c3b-8253a861af54\") " pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-t72zp" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.893320 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2d5ffd79-67fa-4dd0-81a7-b44156946b86-cert\") pod \"openstack-baremetal-operator-controller-manager-659bb84579xswrt\" (UID: \"2d5ffd79-67fa-4dd0-81a7-b44156946b86\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-659bb84579xswrt" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.893390 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-684t7\" (UniqueName: \"kubernetes.io/projected/2d5ffd79-67fa-4dd0-81a7-b44156946b86-kube-api-access-684t7\") pod \"openstack-baremetal-operator-controller-manager-659bb84579xswrt\" (UID: \"2d5ffd79-67fa-4dd0-81a7-b44156946b86\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-659bb84579xswrt" Oct 01 09:49:37 crc kubenswrapper[4787]: E1001 09:49:37.901951 4787 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 01 09:49:37 crc kubenswrapper[4787]: E1001 09:49:37.902060 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2d5ffd79-67fa-4dd0-81a7-b44156946b86-cert podName:2d5ffd79-67fa-4dd0-81a7-b44156946b86 nodeName:}" failed. No retries permitted until 2025-10-01 09:49:38.402036223 +0000 UTC m=+810.517180380 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2d5ffd79-67fa-4dd0-81a7-b44156946b86-cert") pod "openstack-baremetal-operator-controller-manager-659bb84579xswrt" (UID: "2d5ffd79-67fa-4dd0-81a7-b44156946b86") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.933989 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-684t7\" (UniqueName: \"kubernetes.io/projected/2d5ffd79-67fa-4dd0-81a7-b44156946b86-kube-api-access-684t7\") pod \"openstack-baremetal-operator-controller-manager-659bb84579xswrt\" (UID: \"2d5ffd79-67fa-4dd0-81a7-b44156946b86\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-659bb84579xswrt" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.935584 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9e03d54f-d462-4894-9c3b-8253a861af54-cert\") pod \"infra-operator-controller-manager-5c8fdc4d5c-t72zp\" (UID: \"9e03d54f-d462-4894-9c3b-8253a861af54\") " pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-t72zp" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.943620 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-84c745747f-8qtcw"] Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.945752 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-84c745747f-8qtcw" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.957977 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-nbxwx" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.961506 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-jwjcm" Oct 01 09:49:37 crc kubenswrapper[4787]: I1001 09:49:37.977418 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-659bb84579xswrt"] Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:37.999451 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-84c745747f-8qtcw"] Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.018604 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-689b4f76c9-xvrsn"] Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.027241 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-689b4f76c9-xvrsn" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.031315 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-9tdp2" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.032419 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-rjlfq" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.039129 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-75f8d67d86-xjbwt" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.061767 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-5qz2m" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.062033 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-689b4f76c9-xvrsn"] Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.067265 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-598c4c8547-4zm2q"] Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.068689 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-4zm2q" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.086205 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-598c4c8547-4zm2q"] Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.087805 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-vdvwz" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.096243 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-cb66d6b59-fqgnk"] Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.097865 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-fqgnk" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.098789 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5sv6\" (UniqueName: \"kubernetes.io/projected/57f9bf14-0074-43ba-882d-a9102b047a3b-kube-api-access-f5sv6\") pod \"ovn-operator-controller-manager-84c745747f-8qtcw\" (UID: \"57f9bf14-0074-43ba-882d-a9102b047a3b\") " pod="openstack-operators/ovn-operator-controller-manager-84c745747f-8qtcw" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.107770 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-nf4f4" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.119434 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-cbdf6dc66-z2gzf"] Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.121114 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-z2gzf" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.132376 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-cbdf6dc66-z2gzf"] Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.139054 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-cb66d6b59-fqgnk"] Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.143350 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-zzwnz" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.182806 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-68d7bc5569-4mgdw"] Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.195497 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-4mgdw" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.197437 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-t72zp" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.197890 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-68d7bc5569-4mgdw"] Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.201607 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5sv6\" (UniqueName: \"kubernetes.io/projected/57f9bf14-0074-43ba-882d-a9102b047a3b-kube-api-access-f5sv6\") pod \"ovn-operator-controller-manager-84c745747f-8qtcw\" (UID: \"57f9bf14-0074-43ba-882d-a9102b047a3b\") " pod="openstack-operators/ovn-operator-controller-manager-84c745747f-8qtcw" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.201646 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26mgc\" (UniqueName: \"kubernetes.io/projected/5024aa9b-39fc-4e58-96fe-c66d054e3d5a-kube-api-access-26mgc\") pod \"placement-operator-controller-manager-598c4c8547-4zm2q\" (UID: \"5024aa9b-39fc-4e58-96fe-c66d054e3d5a\") " pod="openstack-operators/placement-operator-controller-manager-598c4c8547-4zm2q" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.201706 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqv4h\" (UniqueName: \"kubernetes.io/projected/7139ac70-d405-449b-924e-01fd94e1375a-kube-api-access-nqv4h\") pod \"telemetry-operator-controller-manager-cb66d6b59-fqgnk\" (UID: \"7139ac70-d405-449b-924e-01fd94e1375a\") " pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-fqgnk" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.201733 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fx6r5\" (UniqueName: \"kubernetes.io/projected/a74d851b-a805-4138-9bf4-1a4546a2f30a-kube-api-access-fx6r5\") pod \"test-operator-controller-manager-cbdf6dc66-z2gzf\" (UID: \"a74d851b-a805-4138-9bf4-1a4546a2f30a\") " pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-z2gzf" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.201771 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjppw\" (UniqueName: \"kubernetes.io/projected/ed13cb34-e846-41ab-aff7-8e5d4acd2e6b-kube-api-access-rjppw\") pod \"swift-operator-controller-manager-689b4f76c9-xvrsn\" (UID: \"ed13cb34-e846-41ab-aff7-8e5d4acd2e6b\") " pod="openstack-operators/swift-operator-controller-manager-689b4f76c9-xvrsn" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.210056 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-m72cb" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.253698 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5sv6\" (UniqueName: \"kubernetes.io/projected/57f9bf14-0074-43ba-882d-a9102b047a3b-kube-api-access-f5sv6\") pod \"ovn-operator-controller-manager-84c745747f-8qtcw\" (UID: \"57f9bf14-0074-43ba-882d-a9102b047a3b\") " pod="openstack-operators/ovn-operator-controller-manager-84c745747f-8qtcw" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.254505 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6c7b6bcb7c-bxglr"] Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.258119 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6c7b6bcb7c-bxglr" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.262002 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-bf9jw" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.262525 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.268476 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6c7b6bcb7c-bxglr"] Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.273395 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xvd4r"] Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.274624 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xvd4r" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.277528 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xvd4r"] Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.282409 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-7qkgx" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.303127 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjppw\" (UniqueName: \"kubernetes.io/projected/ed13cb34-e846-41ab-aff7-8e5d4acd2e6b-kube-api-access-rjppw\") pod \"swift-operator-controller-manager-689b4f76c9-xvrsn\" (UID: \"ed13cb34-e846-41ab-aff7-8e5d4acd2e6b\") " pod="openstack-operators/swift-operator-controller-manager-689b4f76c9-xvrsn" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.303207 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rc2w\" (UniqueName: \"kubernetes.io/projected/fb776e43-5f95-44f0-96c0-945b2681d8c1-kube-api-access-9rc2w\") pod \"watcher-operator-controller-manager-68d7bc5569-4mgdw\" (UID: \"fb776e43-5f95-44f0-96c0-945b2681d8c1\") " pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-4mgdw" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.303255 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26mgc\" (UniqueName: \"kubernetes.io/projected/5024aa9b-39fc-4e58-96fe-c66d054e3d5a-kube-api-access-26mgc\") pod \"placement-operator-controller-manager-598c4c8547-4zm2q\" (UID: \"5024aa9b-39fc-4e58-96fe-c66d054e3d5a\") " pod="openstack-operators/placement-operator-controller-manager-598c4c8547-4zm2q" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.303324 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqv4h\" (UniqueName: \"kubernetes.io/projected/7139ac70-d405-449b-924e-01fd94e1375a-kube-api-access-nqv4h\") pod \"telemetry-operator-controller-manager-cb66d6b59-fqgnk\" (UID: \"7139ac70-d405-449b-924e-01fd94e1375a\") " pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-fqgnk" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.303356 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fx6r5\" (UniqueName: \"kubernetes.io/projected/a74d851b-a805-4138-9bf4-1a4546a2f30a-kube-api-access-fx6r5\") pod \"test-operator-controller-manager-cbdf6dc66-z2gzf\" (UID: \"a74d851b-a805-4138-9bf4-1a4546a2f30a\") " pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-z2gzf" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.307198 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-84c745747f-8qtcw" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.316213 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-f7f98cb69-wqs6n"] Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.342208 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26mgc\" (UniqueName: \"kubernetes.io/projected/5024aa9b-39fc-4e58-96fe-c66d054e3d5a-kube-api-access-26mgc\") pod \"placement-operator-controller-manager-598c4c8547-4zm2q\" (UID: \"5024aa9b-39fc-4e58-96fe-c66d054e3d5a\") " pod="openstack-operators/placement-operator-controller-manager-598c4c8547-4zm2q" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.343386 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqv4h\" (UniqueName: \"kubernetes.io/projected/7139ac70-d405-449b-924e-01fd94e1375a-kube-api-access-nqv4h\") pod \"telemetry-operator-controller-manager-cb66d6b59-fqgnk\" (UID: \"7139ac70-d405-449b-924e-01fd94e1375a\") " pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-fqgnk" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.345488 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjppw\" (UniqueName: \"kubernetes.io/projected/ed13cb34-e846-41ab-aff7-8e5d4acd2e6b-kube-api-access-rjppw\") pod \"swift-operator-controller-manager-689b4f76c9-xvrsn\" (UID: \"ed13cb34-e846-41ab-aff7-8e5d4acd2e6b\") " pod="openstack-operators/swift-operator-controller-manager-689b4f76c9-xvrsn" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.346420 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fx6r5\" (UniqueName: \"kubernetes.io/projected/a74d851b-a805-4138-9bf4-1a4546a2f30a-kube-api-access-fx6r5\") pod \"test-operator-controller-manager-cbdf6dc66-z2gzf\" (UID: \"a74d851b-a805-4138-9bf4-1a4546a2f30a\") " pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-z2gzf" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.374205 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-689b4f76c9-xvrsn" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.404650 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6gjf\" (UniqueName: \"kubernetes.io/projected/f31cb2c2-eabc-4bca-88c5-2aa194fb13b5-kube-api-access-f6gjf\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-xvd4r\" (UID: \"f31cb2c2-eabc-4bca-88c5-2aa194fb13b5\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xvd4r" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.405335 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/40d23171-497b-47d9-96f4-51f32d7683a1-cert\") pod \"openstack-operator-controller-manager-6c7b6bcb7c-bxglr\" (UID: \"40d23171-497b-47d9-96f4-51f32d7683a1\") " pod="openstack-operators/openstack-operator-controller-manager-6c7b6bcb7c-bxglr" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.405387 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rc2w\" (UniqueName: \"kubernetes.io/projected/fb776e43-5f95-44f0-96c0-945b2681d8c1-kube-api-access-9rc2w\") pod \"watcher-operator-controller-manager-68d7bc5569-4mgdw\" (UID: \"fb776e43-5f95-44f0-96c0-945b2681d8c1\") " pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-4mgdw" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.405410 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knhxm\" (UniqueName: \"kubernetes.io/projected/40d23171-497b-47d9-96f4-51f32d7683a1-kube-api-access-knhxm\") pod \"openstack-operator-controller-manager-6c7b6bcb7c-bxglr\" (UID: \"40d23171-497b-47d9-96f4-51f32d7683a1\") " pod="openstack-operators/openstack-operator-controller-manager-6c7b6bcb7c-bxglr" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.405453 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2d5ffd79-67fa-4dd0-81a7-b44156946b86-cert\") pod \"openstack-baremetal-operator-controller-manager-659bb84579xswrt\" (UID: \"2d5ffd79-67fa-4dd0-81a7-b44156946b86\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-659bb84579xswrt" Oct 01 09:49:38 crc kubenswrapper[4787]: W1001 09:49:38.406243 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod26b292ae_dcb2_41f6_86b5_2eddfecdf356.slice/crio-4e79c76be0f84d458690b5b6ed8c6bf449e27e0c1ac0967045daf2e9f4b8cc52 WatchSource:0}: Error finding container 4e79c76be0f84d458690b5b6ed8c6bf449e27e0c1ac0967045daf2e9f4b8cc52: Status 404 returned error can't find the container with id 4e79c76be0f84d458690b5b6ed8c6bf449e27e0c1ac0967045daf2e9f4b8cc52 Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.440980 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2d5ffd79-67fa-4dd0-81a7-b44156946b86-cert\") pod \"openstack-baremetal-operator-controller-manager-659bb84579xswrt\" (UID: \"2d5ffd79-67fa-4dd0-81a7-b44156946b86\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-659bb84579xswrt" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.444304 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-4zm2q" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.445244 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rc2w\" (UniqueName: \"kubernetes.io/projected/fb776e43-5f95-44f0-96c0-945b2681d8c1-kube-api-access-9rc2w\") pod \"watcher-operator-controller-manager-68d7bc5569-4mgdw\" (UID: \"fb776e43-5f95-44f0-96c0-945b2681d8c1\") " pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-4mgdw" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.499463 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-z2gzf" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.509546 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6gjf\" (UniqueName: \"kubernetes.io/projected/f31cb2c2-eabc-4bca-88c5-2aa194fb13b5-kube-api-access-f6gjf\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-xvd4r\" (UID: \"f31cb2c2-eabc-4bca-88c5-2aa194fb13b5\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xvd4r" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.509646 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/40d23171-497b-47d9-96f4-51f32d7683a1-cert\") pod \"openstack-operator-controller-manager-6c7b6bcb7c-bxglr\" (UID: \"40d23171-497b-47d9-96f4-51f32d7683a1\") " pod="openstack-operators/openstack-operator-controller-manager-6c7b6bcb7c-bxglr" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.509698 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knhxm\" (UniqueName: \"kubernetes.io/projected/40d23171-497b-47d9-96f4-51f32d7683a1-kube-api-access-knhxm\") pod \"openstack-operator-controller-manager-6c7b6bcb7c-bxglr\" (UID: \"40d23171-497b-47d9-96f4-51f32d7683a1\") " pod="openstack-operators/openstack-operator-controller-manager-6c7b6bcb7c-bxglr" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.511607 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-fqgnk" Oct 01 09:49:38 crc kubenswrapper[4787]: E1001 09:49:38.511784 4787 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 01 09:49:38 crc kubenswrapper[4787]: E1001 09:49:38.511846 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/40d23171-497b-47d9-96f4-51f32d7683a1-cert podName:40d23171-497b-47d9-96f4-51f32d7683a1 nodeName:}" failed. No retries permitted until 2025-10-01 09:49:39.011823664 +0000 UTC m=+811.126967821 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/40d23171-497b-47d9-96f4-51f32d7683a1-cert") pod "openstack-operator-controller-manager-6c7b6bcb7c-bxglr" (UID: "40d23171-497b-47d9-96f4-51f32d7683a1") : secret "webhook-server-cert" not found Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.516154 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-659bb84579xswrt" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.516198 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859cd486d-6h8cm"] Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.542926 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6gjf\" (UniqueName: \"kubernetes.io/projected/f31cb2c2-eabc-4bca-88c5-2aa194fb13b5-kube-api-access-f6gjf\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-xvd4r\" (UID: \"f31cb2c2-eabc-4bca-88c5-2aa194fb13b5\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xvd4r" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.543509 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knhxm\" (UniqueName: \"kubernetes.io/projected/40d23171-497b-47d9-96f4-51f32d7683a1-kube-api-access-knhxm\") pod \"openstack-operator-controller-manager-6c7b6bcb7c-bxglr\" (UID: \"40d23171-497b-47d9-96f4-51f32d7683a1\") " pod="openstack-operators/openstack-operator-controller-manager-6c7b6bcb7c-bxglr" Oct 01 09:49:38 crc kubenswrapper[4787]: W1001 09:49:38.596757 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod24354b04_0daa_4fd4_8d38_b21f64bc2cfe.slice/crio-708cf11e20e0ebe770a2fae5ffe0914bfa6128c1d14326a6eea0330afab92588 WatchSource:0}: Error finding container 708cf11e20e0ebe770a2fae5ffe0914bfa6128c1d14326a6eea0330afab92588: Status 404 returned error can't find the container with id 708cf11e20e0ebe770a2fae5ffe0914bfa6128c1d14326a6eea0330afab92588 Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.631229 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-4mgdw" Oct 01 09:49:38 crc kubenswrapper[4787]: I1001 09:49:38.710895 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xvd4r" Oct 01 09:49:39 crc kubenswrapper[4787]: I1001 09:49:39.028900 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/40d23171-497b-47d9-96f4-51f32d7683a1-cert\") pod \"openstack-operator-controller-manager-6c7b6bcb7c-bxglr\" (UID: \"40d23171-497b-47d9-96f4-51f32d7683a1\") " pod="openstack-operators/openstack-operator-controller-manager-6c7b6bcb7c-bxglr" Oct 01 09:49:39 crc kubenswrapper[4787]: I1001 09:49:39.037358 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/40d23171-497b-47d9-96f4-51f32d7683a1-cert\") pod \"openstack-operator-controller-manager-6c7b6bcb7c-bxglr\" (UID: \"40d23171-497b-47d9-96f4-51f32d7683a1\") " pod="openstack-operators/openstack-operator-controller-manager-6c7b6bcb7c-bxglr" Oct 01 09:49:39 crc kubenswrapper[4787]: I1001 09:49:39.053254 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5b4fc86755-lbbxd"] Oct 01 09:49:39 crc kubenswrapper[4787]: I1001 09:49:39.070598 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-54fbbfcd44-4jlcz"] Oct 01 09:49:39 crc kubenswrapper[4787]: W1001 09:49:39.084546 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb1a0ae18_3c6d_473a_a660_5ff59f6a87bf.slice/crio-467838bc03f18d31cf962fe8e797b5514593860c5ed68ae7ee54bd9ce39750a1 WatchSource:0}: Error finding container 467838bc03f18d31cf962fe8e797b5514593860c5ed68ae7ee54bd9ce39750a1: Status 404 returned error can't find the container with id 467838bc03f18d31cf962fe8e797b5514593860c5ed68ae7ee54bd9ce39750a1 Oct 01 09:49:39 crc kubenswrapper[4787]: I1001 09:49:39.084650 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-8bc4775b5-x2v89"] Oct 01 09:49:39 crc kubenswrapper[4787]: W1001 09:49:39.088611 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8b59a5e1_4f45_4bcd_b9ce_63bae8b5ed57.slice/crio-b0b0d22bd0016a1f361bb09236e550b7dea6114be3c458226ed6a1393291327b WatchSource:0}: Error finding container b0b0d22bd0016a1f361bb09236e550b7dea6114be3c458226ed6a1393291327b: Status 404 returned error can't find the container with id b0b0d22bd0016a1f361bb09236e550b7dea6114be3c458226ed6a1393291327b Oct 01 09:49:39 crc kubenswrapper[4787]: I1001 09:49:39.100891 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-77fb7bcf5b-2pxzs"] Oct 01 09:49:39 crc kubenswrapper[4787]: I1001 09:49:39.222778 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-x2v89" event={"ID":"b4902328-fb90-40f9-a84c-5a0359d699d4","Type":"ContainerStarted","Data":"1212c988ddfd10543a0f117cd461d897ff3caffa4ac29dbc68eb128c16fd5f1b"} Oct 01 09:49:39 crc kubenswrapper[4787]: I1001 09:49:39.234704 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859cd486d-6h8cm" event={"ID":"24354b04-0daa-4fd4-8d38-b21f64bc2cfe","Type":"ContainerStarted","Data":"708cf11e20e0ebe770a2fae5ffe0914bfa6128c1d14326a6eea0330afab92588"} Oct 01 09:49:39 crc kubenswrapper[4787]: I1001 09:49:39.238292 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-2pxzs" event={"ID":"77e07ae2-1dae-4570-84b8-c137e86ac628","Type":"ContainerStarted","Data":"78d9230887d641fbac1226d3ce0533b4386a0b978448ba6b31d79f2b684c6d89"} Oct 01 09:49:39 crc kubenswrapper[4787]: I1001 09:49:39.239418 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-wqs6n" event={"ID":"26b292ae-dcb2-41f6-86b5-2eddfecdf356","Type":"ContainerStarted","Data":"4e79c76be0f84d458690b5b6ed8c6bf449e27e0c1ac0967045daf2e9f4b8cc52"} Oct 01 09:49:39 crc kubenswrapper[4787]: I1001 09:49:39.240246 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-54fbbfcd44-4jlcz" event={"ID":"8b59a5e1-4f45-4bcd-b9ce-63bae8b5ed57","Type":"ContainerStarted","Data":"b0b0d22bd0016a1f361bb09236e550b7dea6114be3c458226ed6a1393291327b"} Oct 01 09:49:39 crc kubenswrapper[4787]: I1001 09:49:39.240907 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-lbbxd" event={"ID":"b1a0ae18-3c6d-473a-a660-5ff59f6a87bf","Type":"ContainerStarted","Data":"467838bc03f18d31cf962fe8e797b5514593860c5ed68ae7ee54bd9ce39750a1"} Oct 01 09:49:39 crc kubenswrapper[4787]: I1001 09:49:39.245050 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-679b4759bb-qt9bq"] Oct 01 09:49:39 crc kubenswrapper[4787]: I1001 09:49:39.249635 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5f45cd594f-zrtv2"] Oct 01 09:49:39 crc kubenswrapper[4787]: W1001 09:49:39.255164 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podae2ba2f0_8422_4a03_b64f_d335f76c5532.slice/crio-4f39076ff1fce73f4b73d6cd7e16dff45dc6f506f63a757c672ff4817e6222ee WatchSource:0}: Error finding container 4f39076ff1fce73f4b73d6cd7e16dff45dc6f506f63a757c672ff4817e6222ee: Status 404 returned error can't find the container with id 4f39076ff1fce73f4b73d6cd7e16dff45dc6f506f63a757c672ff4817e6222ee Oct 01 09:49:39 crc kubenswrapper[4787]: W1001 09:49:39.258455 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f52292a_2603_4524_b3fc_8d7b003f95f8.slice/crio-ee50cd5c458f4ac9f4356b8a36de33d691456ac31ef58772f014be8db21ff398 WatchSource:0}: Error finding container ee50cd5c458f4ac9f4356b8a36de33d691456ac31ef58772f014be8db21ff398: Status 404 returned error can't find the container with id ee50cd5c458f4ac9f4356b8a36de33d691456ac31ef58772f014be8db21ff398 Oct 01 09:49:39 crc kubenswrapper[4787]: I1001 09:49:39.263872 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6c7b6bcb7c-bxglr" Oct 01 09:49:39 crc kubenswrapper[4787]: I1001 09:49:39.596561 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-75f8d67d86-xjbwt"] Oct 01 09:49:39 crc kubenswrapper[4787]: I1001 09:49:39.611274 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-nbxwx"] Oct 01 09:49:39 crc kubenswrapper[4787]: I1001 09:49:39.629349 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-84c745747f-8qtcw"] Oct 01 09:49:39 crc kubenswrapper[4787]: W1001 09:49:39.642379 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod57f9bf14_0074_43ba_882d_a9102b047a3b.slice/crio-52ca4e9b6e18eddc17e10992c0fafcdd88599e77093350a0d5f442651ea3d28b WatchSource:0}: Error finding container 52ca4e9b6e18eddc17e10992c0fafcdd88599e77093350a0d5f442651ea3d28b: Status 404 returned error can't find the container with id 52ca4e9b6e18eddc17e10992c0fafcdd88599e77093350a0d5f442651ea3d28b Oct 01 09:49:39 crc kubenswrapper[4787]: I1001 09:49:39.646603 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67bf5bb885-qmr4v"] Oct 01 09:49:39 crc kubenswrapper[4787]: I1001 09:49:39.652515 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-68d7bc5569-4mgdw"] Oct 01 09:49:39 crc kubenswrapper[4787]: W1001 09:49:39.654154 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2dfc2636_198f_4b51_b07e_40ff76c1091c.slice/crio-07f4b71c6403b3925ce9e5af6b4679b5b46c8f8eb9a4c4423a81b96e4e4da61a WatchSource:0}: Error finding container 07f4b71c6403b3925ce9e5af6b4679b5b46c8f8eb9a4c4423a81b96e4e4da61a: Status 404 returned error can't find the container with id 07f4b71c6403b3925ce9e5af6b4679b5b46c8f8eb9a4c4423a81b96e4e4da61a Oct 01 09:49:39 crc kubenswrapper[4787]: W1001 09:49:39.656359 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poded13cb34_e846_41ab_aff7_8e5d4acd2e6b.slice/crio-d968cf8f9a7012b9a3f5f0c5bd2879c94f3aa2196644be802dec25ba811d6a73 WatchSource:0}: Error finding container d968cf8f9a7012b9a3f5f0c5bd2879c94f3aa2196644be802dec25ba811d6a73: Status 404 returned error can't find the container with id d968cf8f9a7012b9a3f5f0c5bd2879c94f3aa2196644be802dec25ba811d6a73 Oct 01 09:49:39 crc kubenswrapper[4787]: I1001 09:49:39.656507 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-t72zp"] Oct 01 09:49:39 crc kubenswrapper[4787]: I1001 09:49:39.660299 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-b7cf8cb5f-5qz2m"] Oct 01 09:49:39 crc kubenswrapper[4787]: W1001 09:49:39.661865 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb54a691a_cd96_4ae4_a9ce_65662f58cd60.slice/crio-bb73969164798f0ed06f5ef083e73726130f1c68d100bd5b674848c6940e3586 WatchSource:0}: Error finding container bb73969164798f0ed06f5ef083e73726130f1c68d100bd5b674848c6940e3586: Status 404 returned error can't find the container with id bb73969164798f0ed06f5ef083e73726130f1c68d100bd5b674848c6940e3586 Oct 01 09:49:39 crc kubenswrapper[4787]: W1001 09:49:39.662687 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfb776e43_5f95_44f0_96c0_945b2681d8c1.slice/crio-6551bdf6f44aab116783d132d4bde7dc34e21b5cc9ac9c634985fd392cef3cf3 WatchSource:0}: Error finding container 6551bdf6f44aab116783d132d4bde7dc34e21b5cc9ac9c634985fd392cef3cf3: Status 404 returned error can't find the container with id 6551bdf6f44aab116783d132d4bde7dc34e21b5cc9ac9c634985fd392cef3cf3 Oct 01 09:49:39 crc kubenswrapper[4787]: I1001 09:49:39.663775 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-689b4f76c9-xvrsn"] Oct 01 09:49:39 crc kubenswrapper[4787]: E1001 09:49:39.665942 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:15d7b5a365350a831ca59d984df67fadeccf89d599e487a7597b105afb82ce4a,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-442mm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-67bf5bb885-qmr4v_openstack-operators(b54a691a-cd96-4ae4-a9ce-65662f58cd60): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 01 09:49:39 crc kubenswrapper[4787]: E1001 09:49:39.669132 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:09c2f519ea218f6038b7be039b8e6ac33ee93b217b9be0d2d18a5e7f94faae06,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9rc2w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-68d7bc5569-4mgdw_openstack-operators(fb776e43-5f95-44f0-96c0-945b2681d8c1): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 01 09:49:39 crc kubenswrapper[4787]: I1001 09:49:39.672826 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-59d7dc95cf-9tdp2"] Oct 01 09:49:39 crc kubenswrapper[4787]: E1001 09:49:39.682298 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:23fcec0642cbd40af10bca0c5d4e538662d21eda98d6dfec37c38b4d7a47191a,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6tdhg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-59d7dc95cf-9tdp2_openstack-operators(88d93cbc-27c2-47be-a882-6607925d4391): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 01 09:49:39 crc kubenswrapper[4787]: E1001 09:49:39.877523 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-qmr4v" podUID="b54a691a-cd96-4ae4-a9ce-65662f58cd60" Oct 01 09:49:39 crc kubenswrapper[4787]: E1001 09:49:39.891906 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-4mgdw" podUID="fb776e43-5f95-44f0-96c0-945b2681d8c1" Oct 01 09:49:39 crc kubenswrapper[4787]: E1001 09:49:39.908491 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-9tdp2" podUID="88d93cbc-27c2-47be-a882-6607925d4391" Oct 01 09:49:39 crc kubenswrapper[4787]: I1001 09:49:39.935087 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-cb66d6b59-fqgnk"] Oct 01 09:49:39 crc kubenswrapper[4787]: I1001 09:49:39.938825 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xvd4r"] Oct 01 09:49:39 crc kubenswrapper[4787]: I1001 09:49:39.952985 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-598c4c8547-4zm2q"] Oct 01 09:49:39 crc kubenswrapper[4787]: I1001 09:49:39.961065 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-cbdf6dc66-z2gzf"] Oct 01 09:49:39 crc kubenswrapper[4787]: I1001 09:49:39.986715 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-659bb84579xswrt"] Oct 01 09:49:40 crc kubenswrapper[4787]: E1001 09:49:40.011358 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-f6gjf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-xvd4r_openstack-operators(f31cb2c2-eabc-4bca-88c5-2aa194fb13b5): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 01 09:49:40 crc kubenswrapper[4787]: E1001 09:49:40.012592 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xvd4r" podUID="f31cb2c2-eabc-4bca-88c5-2aa194fb13b5" Oct 01 09:49:40 crc kubenswrapper[4787]: E1001 09:49:40.027783 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:f61fdfbfd12027ce6b4e7ad553ec0582f080de0cfb472de6dc04ad3078bb17e3,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fx6r5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-cbdf6dc66-z2gzf_openstack-operators(a74d851b-a805-4138-9bf4-1a4546a2f30a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 01 09:49:40 crc kubenswrapper[4787]: E1001 09:49:40.028235 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:e7cfed051c1cf801e651fd4035070e38698039f284ac0b2a0332769fdbb4a9c8,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent@sha256:9739588b6480acdeada79842182c7e8507dc4f3669be8330591460ffd44cdcec,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:9774f19d7a63d6f516afa701fb5f031674ad537e595049bbc57817356c7642fe,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api@sha256:29c8cd4f2d853f512e2ecd44f522f28c3aac046a72733365aa5e91667041d62e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator@sha256:ed896681f0d9720f56bbcb0b7a4f3626ed397e89af919604ca68b42b7b598859,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener@sha256:712e1c932a90ef5e3c3ee5d5aea591a377da8c4af604ebd8ec399869a61dfbef,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier@sha256:10fd8489a5bf6f1d781e9226de68356132db78b62269e69d632748cb08fae725,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24@sha256:e91d58021b54c46883595ff66be65882de54abdb3be2ca53c4162b20d18b5f48,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:73fd28af83ea96cc920d26dba6105ee59f0824234527949884e6ca55b71d7533,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener@sha256:8b3a90516ba0695cf3198a7b101da770c30c8100cb79f8088b5729e6a50ddd6d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker@sha256:6d42bcf65422d2de9cd807feb3e8b005de10084b4b8eb340c8a9045644ae7aaa,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:32a25ac44706b73bff04a89514177b1efd675f0442b295e225f0020555ca6350,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute@sha256:b19043eac7c653e00da8da9418ae378fdd29698adb1adb4bf5ae7cfc03ba5538,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi@sha256:c486e00b36ea7698d6a4cd9048a759bad5a8286e4949bbd1f82c3ddb70600b9b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter@sha256:7211a617ec657701ca819aa0ba28e1d5750f5bf2c1391b755cc4a48cc360b0fa,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification@sha256:ef2727f0300fbf3bf15d8ddc409d0fd63e4aac9dd64c86459bd6ff64fc6b9534,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core@sha256:09b5017c95d7697e66b9c64846bc48ef5826a009cba89b956ec54561e5f4a2d1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:e318869f706836a0c74c0ad55aab277b1bb7fae0555ae0f03cb28b379b9ce695,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup@sha256:329aac65ba00c3cf43bb1d5fac8818752f01de90b47719e2a84db4e2fe083292,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler@sha256:6ce73885ac1ee7c69468efc448eff5deae46502812c5e3d099f771e1cc03345f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume@sha256:282cc0fcdbb8a688dd62a2499480aae4a36b620f2160d51e6c8269e6cc32d5fc,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api@sha256:d98c0c9d3bdd84daf4b98d45b8bbe2e67a633491897dda7167664a5fa1f0f26e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9@sha256:4ad1d36fe1c8992e43910fc2d566b991fd73f9b82b1ab860c66858448ff82c00,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central@sha256:92789eab1b8a91807a5e898cb63478d125ae539eafe63c96049100c6ddeadb04,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns@sha256:ee9832268e0df5d62c50c5ce171e9ef72a035aa74c718cfbf482e34426d8d15e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer@sha256:07b4f96f24f32224c13613f85173f9fcc3092b8797ffa47519403d124bfe4c15,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound@sha256:3a873c95bcb7ae8bd24ff1eb5fe89ac5272a41a3345a7b41d55419b5d66b70e7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker@sha256:388dbae2f1aae2720e919cc24d10cd577b73b4e4ef7abdc34287bcb8d27ff98f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr@sha256:d4c1b2496868da3dcca9f4bda0834fcc58d23c21d8ce3c42a68205d02039c487,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid@sha256:c4414cc2680fb1bacbf99261f759f4ef7401fb2e4953140270bffdab8e002f22,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler@sha256:581b65b646301e0fcb07582150ba63438f1353a85bf9acf1eb2acb4ce71c58bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron@sha256:b9b950a656f1456b3143872c492b0987bf4a9e23bc7c59d843cf50099667b368,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd@sha256:afd5d6822b86ea0930b2011fede834bb24495995d7baac03363ab61d89f07a22,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent@sha256:665d7a25dfc959ec5448d5ba6b430792ebde1be1580ea6809e9b3b4f94184b3f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn@sha256:499c6d82390ee2dbb91628d2e42671406372fb603d697685a04145cf6dd8d0ab,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent@sha256:da2736bc98bfe340e86234523d4c00220f6f79add271900981cf4ad9f4c5ee51,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent@sha256:4df8dad8a5fb4805a0424cbc0b8df666b9a06b76c64f26e186f3b9e8efe6cd95,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent@sha256:65c16453b5b7bb113646ffce0be26138e89eecbf6dd1582cdfe76af7f5dc62cf,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api@sha256:bdfed2a176a064bf70082602a1f319eace2d9003ff1117b1e48b7f2130840070,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api@sha256:ce968dce2209ec5114772b4b73ed16c0a25988637372f2afbfac080cc6f1e378,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn@sha256:b7823eaacf55280cdf3f1bede4f40bf49fdbf9ba9f3f5ba64b0abedede601c8f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine@sha256:605206d967ffaa20156eb07a645654cd3e0f880bb0eefbb2b5e1e749b169f148,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon@sha256:9470db6caf5102cf37ddb1f137f17b05ef7119f174f4189beb4839ef7f65730c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached@sha256:34e84da4ae7e5d65931cbefcda84fd8fdc93271ec466adf1a9040b67a3af176a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis@sha256:b301b17c31e47733a8a232773427ce3cb50433a3aa09d4a5bd998b1aeb5e5530,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api@sha256:d642c35c0f9d3acf31987c028f1d4d4fdf7b49e1d6cbcd73268c12b3d6e14b86,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor@sha256:922eb0799ab36a91aa95abe52565dc60db807457dbf8c651b30e06b9e8aebcd4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector@sha256:cd01e9605ab513458a6813e38d37fbfde1a91388cc5c00962203dbcbdc285e79,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent@sha256:dd35c22b17730cbca8547ea98459f182939462c8dc3465d21335a377018937de,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe@sha256:0e0e2e48a41d5417f1d6a4407e63d443611b7eacd66e27f561c9eedf3e5a66c5,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent@sha256:735bd24219fdb5f21c31313a5bc685364f45c004fb5e8af634984c147060d4e4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone@sha256:35b5554efae34f2c25a2d274c78bdaecf3d4ce949fa61c692835ee54cdfc6d74,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api@sha256:01b93ab0d87482b9a1fd46706771974743dea1ca74f5fcc3de4a560f7cfc033b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler@sha256:87471fbe3ba77b7115096f4fef8f5a9e1468cbd5bf6060c09785a60f9107a717,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share@sha256:947dcc46173064939cba252d5db34eb6ddd05eb0af7afd762beebe77e9a72c6e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:8498ed720d02ce4e7045f7eb0051b138274cddba9b1e443d11e413da3474d3a3,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils@sha256:2cb054830655a6af5fc6848360618676d24fd9cf15078c0b9855e09d05733eec,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:6276771339c90f342673dcaf7faa8c46e2c0ece62ed5efc4b7d65a095dabe07b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api@sha256:0f5f8f560cd3b4951f7e8e67ef570575435b4c6915658cbb66f32a201776078b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute@sha256:7055e8d7b7d72ce697c6077be14c525c019d186002f04765b90a14c82e01cc7c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor@sha256:d2cd7a21461b4b569d93a63d57761f437cf6bd0847d69a3a65f64d400c7cca6d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy@sha256:432c0c6f36a5e4e4db394771f7dc72f3bf9e5060dc4220f781d3c5050cc17f0d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler@sha256:3ff379a74cc15352bfa25605dbb1a5f4250620e8364bf87ed2f3d5c17e6a8b26,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api@sha256:c67a7bba2fc9351c302369b590473a737bab20d0982d227756fe1fa0bc1c8773,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager@sha256:50c613d159667a26ba4bfb7aebf157b8db8919c815a866438b1d2700231a508e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping@sha256:f3d3d7a7c83926a09714199406bfe8070e6be5055cbfbf00aa37f47e1e5e9bc9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog@sha256:e9b3260907b0e417bb779a7d513a2639734cbbf792e77c61e05e760d06978f4a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker@sha256:1aa6a76e67f2d91ee45472741238b5d4ab53f9bcb94db678c7ae92e1af28899d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient@sha256:80b8547cf5821a4eb5461d1ac14edbc700ef03926268af960bf511647de027af,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_LIGHTSPEED_IMAGE_URL_DEFAULT,Value:quay.io/openstack-lightspeed/rag-content@sha256:5b82cdbfa30e915f97ab6f3726b60582c7b62a819e4aa4e87cf42fc7495b4ef9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather@sha256:bf42dfd2e225818662aa28c4bb23204dc47b2b91127ca0e49b085baa1ea7609d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi@sha256:bd08ffdb4dcfd436200d846d15b2bdcc14122fa43adfea4c0980a087a18f9e3e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller@sha256:2d1e733d24df6ca02636374147f801a0ec1509f8db2f9ad8c739b3f2341815fd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base@sha256:c08ba2a0df4cc18e615b25c329e9c74153709b435c032c38502ec78ba297c5fb,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server@sha256:b6cdafc7722def5b63ef4f00251e10aca93ef82628b21e88925c3d4b49277316,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd@sha256:0a0bbe43e3c266dfeb40a09036f76393dc70377b636724c130a29c434f6d6c82,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server@sha256:7387b628d7cfb3ff349e0df6f11f41ae7fdb0e2d55844944896af02a81ac7cf7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api@sha256:9a3671dee1752ebe3639a0b16de95d29e779f1629d563e0585d65b9792542fc9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:37cc031749b113c35231066ce9f8ce7ccc83e21808ba92ea1981e72bbc42e80f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account@sha256:b2782fe02b1438d68308a5847b0628f0971b5bb8bb0a4d20fe15176fa75bd33f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container@sha256:7118cc3a695fead2a8bab14c8ace018ed7a5ba23ef347bf4ead44219e8467866,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object@sha256:793a836e17b07b0e0a4e8d3177fd04724e1e058fca275ef434abe60a2e444a79,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server@sha256:713d74dc81859344bdcae68a9f7a954146c3e68cfa819518a58cce9e896298c8,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all@sha256:e39be536015777a1b0df8ac863f354046b2b15fee8482abd37d2fa59d8074208,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api@sha256:6ab460c1ec80799499eae55bb8cad0ac3bd3e501c7abe57b665e58921ca88063,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier@sha256:f3d12096a1cd68b1aa837f46a42418ba8a11ca2d18dcb63e5c16d15986f28d4c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine@sha256:700e3619916939c838032c130f0e4a0337a628598ae6a7b752a8e4733bb231e0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-684t7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-659bb84579xswrt_openstack-operators(2d5ffd79-67fa-4dd0-81a7-b44156946b86): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 01 09:49:40 crc kubenswrapper[4787]: I1001 09:49:40.034260 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6c7b6bcb7c-bxglr"] Oct 01 09:49:40 crc kubenswrapper[4787]: I1001 09:49:40.269855 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6c7b6bcb7c-bxglr" event={"ID":"40d23171-497b-47d9-96f4-51f32d7683a1","Type":"ContainerStarted","Data":"01e18924b7d0a144865cf9a4b41a74f44c49de01008185fdbc8e5754cc51b27d"} Oct 01 09:49:40 crc kubenswrapper[4787]: I1001 09:49:40.271369 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-qt9bq" event={"ID":"8f52292a-2603-4524-b3fc-8d7b003f95f8","Type":"ContainerStarted","Data":"ee50cd5c458f4ac9f4356b8a36de33d691456ac31ef58772f014be8db21ff398"} Oct 01 09:49:40 crc kubenswrapper[4787]: I1001 09:49:40.273416 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-z2gzf" event={"ID":"a74d851b-a805-4138-9bf4-1a4546a2f30a","Type":"ContainerStarted","Data":"0ad357d061482e90169b73746e536105c878849a579fc1487ee93275327a3f2c"} Oct 01 09:49:40 crc kubenswrapper[4787]: I1001 09:49:40.277126 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-9tdp2" event={"ID":"88d93cbc-27c2-47be-a882-6607925d4391","Type":"ContainerStarted","Data":"383e038c7f6bf5333f7d7ba559f6b8cb29fbebd3b54ec28eb11093e54f97b9bc"} Oct 01 09:49:40 crc kubenswrapper[4787]: I1001 09:49:40.277158 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-9tdp2" event={"ID":"88d93cbc-27c2-47be-a882-6607925d4391","Type":"ContainerStarted","Data":"059bb2e5b8d3b6b4038de5c9144a30f03ee8b56e8a6b51185aaf6bd2bb50f564"} Oct 01 09:49:40 crc kubenswrapper[4787]: E1001 09:49:40.279261 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:23fcec0642cbd40af10bca0c5d4e538662d21eda98d6dfec37c38b4d7a47191a\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-9tdp2" podUID="88d93cbc-27c2-47be-a882-6607925d4391" Oct 01 09:49:40 crc kubenswrapper[4787]: I1001 09:49:40.295277 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-fqgnk" event={"ID":"7139ac70-d405-449b-924e-01fd94e1375a","Type":"ContainerStarted","Data":"980a8a857baeb539964fccfd718478faf253b521f0f8d66e503974f21232f7ac"} Oct 01 09:49:40 crc kubenswrapper[4787]: E1001 09:49:40.306331 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-z2gzf" podUID="a74d851b-a805-4138-9bf4-1a4546a2f30a" Oct 01 09:49:40 crc kubenswrapper[4787]: I1001 09:49:40.318968 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-qmr4v" event={"ID":"b54a691a-cd96-4ae4-a9ce-65662f58cd60","Type":"ContainerStarted","Data":"496c1043dc8ca4aa51d54d82fd37efa9750fb9c960ce23fab03f1bafd46c8a74"} Oct 01 09:49:40 crc kubenswrapper[4787]: I1001 09:49:40.319049 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-qmr4v" event={"ID":"b54a691a-cd96-4ae4-a9ce-65662f58cd60","Type":"ContainerStarted","Data":"bb73969164798f0ed06f5ef083e73726130f1c68d100bd5b674848c6940e3586"} Oct 01 09:49:40 crc kubenswrapper[4787]: E1001 09:49:40.323695 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:15d7b5a365350a831ca59d984df67fadeccf89d599e487a7597b105afb82ce4a\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-qmr4v" podUID="b54a691a-cd96-4ae4-a9ce-65662f58cd60" Oct 01 09:49:40 crc kubenswrapper[4787]: I1001 09:49:40.329283 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-75f8d67d86-xjbwt" event={"ID":"20e0b9a6-f51b-4925-bffd-8fa00e9a8ec7","Type":"ContainerStarted","Data":"4458e21e8faa3cd6899277e9cdea5e52900989aeb83055b368f4916ee9bb0f18"} Oct 01 09:49:40 crc kubenswrapper[4787]: I1001 09:49:40.330995 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xvd4r" event={"ID":"f31cb2c2-eabc-4bca-88c5-2aa194fb13b5","Type":"ContainerStarted","Data":"a7780e568bd20186f99c77681b303d6d299820bf4a72df20112c7ccb43af5a82"} Oct 01 09:49:40 crc kubenswrapper[4787]: E1001 09:49:40.332245 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xvd4r" podUID="f31cb2c2-eabc-4bca-88c5-2aa194fb13b5" Oct 01 09:49:40 crc kubenswrapper[4787]: I1001 09:49:40.334150 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-5qz2m" event={"ID":"2dfc2636-198f-4b51-b07e-40ff76c1091c","Type":"ContainerStarted","Data":"07f4b71c6403b3925ce9e5af6b4679b5b46c8f8eb9a4c4423a81b96e4e4da61a"} Oct 01 09:49:40 crc kubenswrapper[4787]: I1001 09:49:40.335885 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5f45cd594f-zrtv2" event={"ID":"ae2ba2f0-8422-4a03-b64f-d335f76c5532","Type":"ContainerStarted","Data":"4f39076ff1fce73f4b73d6cd7e16dff45dc6f506f63a757c672ff4817e6222ee"} Oct 01 09:49:40 crc kubenswrapper[4787]: I1001 09:49:40.340816 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-689b4f76c9-xvrsn" event={"ID":"ed13cb34-e846-41ab-aff7-8e5d4acd2e6b","Type":"ContainerStarted","Data":"d968cf8f9a7012b9a3f5f0c5bd2879c94f3aa2196644be802dec25ba811d6a73"} Oct 01 09:49:40 crc kubenswrapper[4787]: I1001 09:49:40.382115 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-4mgdw" event={"ID":"fb776e43-5f95-44f0-96c0-945b2681d8c1","Type":"ContainerStarted","Data":"980bbba60c67df8ce581fbcace01b1cdc14c4c1411e7cd6aac4b4a1a0d37ecb2"} Oct 01 09:49:40 crc kubenswrapper[4787]: I1001 09:49:40.382220 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-4mgdw" event={"ID":"fb776e43-5f95-44f0-96c0-945b2681d8c1","Type":"ContainerStarted","Data":"6551bdf6f44aab116783d132d4bde7dc34e21b5cc9ac9c634985fd392cef3cf3"} Oct 01 09:49:40 crc kubenswrapper[4787]: I1001 09:49:40.384511 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-659bb84579xswrt" event={"ID":"2d5ffd79-67fa-4dd0-81a7-b44156946b86","Type":"ContainerStarted","Data":"2a6c8532d0104a4bd04f573a36e32a96adcb033c2ab0ec5ab463fe0ac9508aa6"} Oct 01 09:49:40 crc kubenswrapper[4787]: E1001 09:49:40.386578 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:09c2f519ea218f6038b7be039b8e6ac33ee93b217b9be0d2d18a5e7f94faae06\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-4mgdw" podUID="fb776e43-5f95-44f0-96c0-945b2681d8c1" Oct 01 09:49:40 crc kubenswrapper[4787]: I1001 09:49:40.386995 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-nbxwx" event={"ID":"c186a6f1-61b9-4ab7-8996-5e1d8439a77c","Type":"ContainerStarted","Data":"a5bce2dd4496f9f2d90a0eac13f5caa086afec17695e6ae8b4e4e51739add346"} Oct 01 09:49:40 crc kubenswrapper[4787]: I1001 09:49:40.390347 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-4zm2q" event={"ID":"5024aa9b-39fc-4e58-96fe-c66d054e3d5a","Type":"ContainerStarted","Data":"7f81446916a4819a960464e4078dea533db23f28a96491deae58d80c2cd49e21"} Oct 01 09:49:40 crc kubenswrapper[4787]: I1001 09:49:40.391981 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-t72zp" event={"ID":"9e03d54f-d462-4894-9c3b-8253a861af54","Type":"ContainerStarted","Data":"0cae6bf6c15fd2f02c37f7be80e2ee68b11dc70ca4677546c8ec82f156d07004"} Oct 01 09:49:40 crc kubenswrapper[4787]: I1001 09:49:40.403251 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-84c745747f-8qtcw" event={"ID":"57f9bf14-0074-43ba-882d-a9102b047a3b","Type":"ContainerStarted","Data":"52ca4e9b6e18eddc17e10992c0fafcdd88599e77093350a0d5f442651ea3d28b"} Oct 01 09:49:40 crc kubenswrapper[4787]: E1001 09:49:40.472590 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-659bb84579xswrt" podUID="2d5ffd79-67fa-4dd0-81a7-b44156946b86" Oct 01 09:49:41 crc kubenswrapper[4787]: I1001 09:49:41.412679 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-z2gzf" event={"ID":"a74d851b-a805-4138-9bf4-1a4546a2f30a","Type":"ContainerStarted","Data":"031e3fb6abe29eb28b79ddc2b3e9d287008b04eb6b5e8116a64534039324f866"} Oct 01 09:49:41 crc kubenswrapper[4787]: E1001 09:49:41.417196 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:f61fdfbfd12027ce6b4e7ad553ec0582f080de0cfb472de6dc04ad3078bb17e3\\\"\"" pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-z2gzf" podUID="a74d851b-a805-4138-9bf4-1a4546a2f30a" Oct 01 09:49:41 crc kubenswrapper[4787]: I1001 09:49:41.417451 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6c7b6bcb7c-bxglr" event={"ID":"40d23171-497b-47d9-96f4-51f32d7683a1","Type":"ContainerStarted","Data":"7c60ef6809292c6a96c858649914ed0c9d1c20f3b38e6e6ede70913f7391fb4b"} Oct 01 09:49:41 crc kubenswrapper[4787]: I1001 09:49:41.417528 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6c7b6bcb7c-bxglr" event={"ID":"40d23171-497b-47d9-96f4-51f32d7683a1","Type":"ContainerStarted","Data":"d1a360208ad6b025972ad142b1b13f6565e206e18b37a125d73df799b1024330"} Oct 01 09:49:41 crc kubenswrapper[4787]: I1001 09:49:41.417607 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-6c7b6bcb7c-bxglr" Oct 01 09:49:41 crc kubenswrapper[4787]: I1001 09:49:41.423695 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-659bb84579xswrt" event={"ID":"2d5ffd79-67fa-4dd0-81a7-b44156946b86","Type":"ContainerStarted","Data":"eaf19172d16f182ad5f9c9fc78ecea15a8b9c96fcc4d23970a2688c4e3159d1c"} Oct 01 09:49:41 crc kubenswrapper[4787]: E1001 09:49:41.426298 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:e7cfed051c1cf801e651fd4035070e38698039f284ac0b2a0332769fdbb4a9c8\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-659bb84579xswrt" podUID="2d5ffd79-67fa-4dd0-81a7-b44156946b86" Oct 01 09:49:41 crc kubenswrapper[4787]: E1001 09:49:41.428945 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:09c2f519ea218f6038b7be039b8e6ac33ee93b217b9be0d2d18a5e7f94faae06\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-4mgdw" podUID="fb776e43-5f95-44f0-96c0-945b2681d8c1" Oct 01 09:49:41 crc kubenswrapper[4787]: E1001 09:49:41.428972 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xvd4r" podUID="f31cb2c2-eabc-4bca-88c5-2aa194fb13b5" Oct 01 09:49:41 crc kubenswrapper[4787]: E1001 09:49:41.429099 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:15d7b5a365350a831ca59d984df67fadeccf89d599e487a7597b105afb82ce4a\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-qmr4v" podUID="b54a691a-cd96-4ae4-a9ce-65662f58cd60" Oct 01 09:49:41 crc kubenswrapper[4787]: E1001 09:49:41.429118 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:23fcec0642cbd40af10bca0c5d4e538662d21eda98d6dfec37c38b4d7a47191a\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-9tdp2" podUID="88d93cbc-27c2-47be-a882-6607925d4391" Oct 01 09:49:41 crc kubenswrapper[4787]: I1001 09:49:41.514477 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-6c7b6bcb7c-bxglr" podStartSLOduration=4.514450001 podStartE2EDuration="4.514450001s" podCreationTimestamp="2025-10-01 09:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:49:41.499195628 +0000 UTC m=+813.614339805" watchObservedRunningTime="2025-10-01 09:49:41.514450001 +0000 UTC m=+813.629594158" Oct 01 09:49:42 crc kubenswrapper[4787]: E1001 09:49:42.436441 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:e7cfed051c1cf801e651fd4035070e38698039f284ac0b2a0332769fdbb4a9c8\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-659bb84579xswrt" podUID="2d5ffd79-67fa-4dd0-81a7-b44156946b86" Oct 01 09:49:42 crc kubenswrapper[4787]: E1001 09:49:42.437379 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:f61fdfbfd12027ce6b4e7ad553ec0582f080de0cfb472de6dc04ad3078bb17e3\\\"\"" pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-z2gzf" podUID="a74d851b-a805-4138-9bf4-1a4546a2f30a" Oct 01 09:49:42 crc kubenswrapper[4787]: I1001 09:49:42.542374 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-828zk" Oct 01 09:49:42 crc kubenswrapper[4787]: I1001 09:49:42.542445 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-828zk" Oct 01 09:49:42 crc kubenswrapper[4787]: I1001 09:49:42.588759 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-828zk" Oct 01 09:49:43 crc kubenswrapper[4787]: I1001 09:49:43.481792 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-828zk" Oct 01 09:49:43 crc kubenswrapper[4787]: I1001 09:49:43.531227 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-828zk"] Oct 01 09:49:45 crc kubenswrapper[4787]: I1001 09:49:45.453516 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-828zk" podUID="ce71de87-f660-4d9e-845c-1f722a75aa54" containerName="registry-server" containerID="cri-o://690a760a08ddfc9c719b917a5782b9472b36455248abaf828731f30bb471b8e4" gracePeriod=2 Oct 01 09:49:46 crc kubenswrapper[4787]: I1001 09:49:46.463226 4787 generic.go:334] "Generic (PLEG): container finished" podID="ce71de87-f660-4d9e-845c-1f722a75aa54" containerID="690a760a08ddfc9c719b917a5782b9472b36455248abaf828731f30bb471b8e4" exitCode=0 Oct 01 09:49:46 crc kubenswrapper[4787]: I1001 09:49:46.463282 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-828zk" event={"ID":"ce71de87-f660-4d9e-845c-1f722a75aa54","Type":"ContainerDied","Data":"690a760a08ddfc9c719b917a5782b9472b36455248abaf828731f30bb471b8e4"} Oct 01 09:49:49 crc kubenswrapper[4787]: I1001 09:49:49.271501 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-6c7b6bcb7c-bxglr" Oct 01 09:49:51 crc kubenswrapper[4787]: E1001 09:49:51.067216 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:ae0bf0aa387864d26d4cbd04d84f6362d53b721348188f096aaa48757419c4f5" Oct 01 09:49:51 crc kubenswrapper[4787]: E1001 09:49:51.067388 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:ae0bf0aa387864d26d4cbd04d84f6362d53b721348188f096aaa48757419c4f5,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n4mvm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-5f45cd594f-zrtv2_openstack-operators(ae2ba2f0-8422-4a03-b64f-d335f76c5532): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 09:49:51 crc kubenswrapper[4787]: I1001 09:49:51.957262 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-828zk" Oct 01 09:49:52 crc kubenswrapper[4787]: I1001 09:49:52.071357 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce71de87-f660-4d9e-845c-1f722a75aa54-utilities\") pod \"ce71de87-f660-4d9e-845c-1f722a75aa54\" (UID: \"ce71de87-f660-4d9e-845c-1f722a75aa54\") " Oct 01 09:49:52 crc kubenswrapper[4787]: I1001 09:49:52.071431 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lhm7t\" (UniqueName: \"kubernetes.io/projected/ce71de87-f660-4d9e-845c-1f722a75aa54-kube-api-access-lhm7t\") pod \"ce71de87-f660-4d9e-845c-1f722a75aa54\" (UID: \"ce71de87-f660-4d9e-845c-1f722a75aa54\") " Oct 01 09:49:52 crc kubenswrapper[4787]: I1001 09:49:52.071484 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce71de87-f660-4d9e-845c-1f722a75aa54-catalog-content\") pod \"ce71de87-f660-4d9e-845c-1f722a75aa54\" (UID: \"ce71de87-f660-4d9e-845c-1f722a75aa54\") " Oct 01 09:49:52 crc kubenswrapper[4787]: E1001 09:49:52.073019 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-5f45cd594f-zrtv2" podUID="ae2ba2f0-8422-4a03-b64f-d335f76c5532" Oct 01 09:49:52 crc kubenswrapper[4787]: I1001 09:49:52.074043 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce71de87-f660-4d9e-845c-1f722a75aa54-utilities" (OuterVolumeSpecName: "utilities") pod "ce71de87-f660-4d9e-845c-1f722a75aa54" (UID: "ce71de87-f660-4d9e-845c-1f722a75aa54"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:49:52 crc kubenswrapper[4787]: I1001 09:49:52.087426 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce71de87-f660-4d9e-845c-1f722a75aa54-kube-api-access-lhm7t" (OuterVolumeSpecName: "kube-api-access-lhm7t") pod "ce71de87-f660-4d9e-845c-1f722a75aa54" (UID: "ce71de87-f660-4d9e-845c-1f722a75aa54"). InnerVolumeSpecName "kube-api-access-lhm7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:49:52 crc kubenswrapper[4787]: I1001 09:49:52.093849 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce71de87-f660-4d9e-845c-1f722a75aa54-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ce71de87-f660-4d9e-845c-1f722a75aa54" (UID: "ce71de87-f660-4d9e-845c-1f722a75aa54"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:49:52 crc kubenswrapper[4787]: I1001 09:49:52.173966 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce71de87-f660-4d9e-845c-1f722a75aa54-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:49:52 crc kubenswrapper[4787]: I1001 09:49:52.174003 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhm7t\" (UniqueName: \"kubernetes.io/projected/ce71de87-f660-4d9e-845c-1f722a75aa54-kube-api-access-lhm7t\") on node \"crc\" DevicePath \"\"" Oct 01 09:49:52 crc kubenswrapper[4787]: I1001 09:49:52.174016 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce71de87-f660-4d9e-845c-1f722a75aa54-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:49:52 crc kubenswrapper[4787]: I1001 09:49:52.549412 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-t72zp" event={"ID":"9e03d54f-d462-4894-9c3b-8253a861af54","Type":"ContainerStarted","Data":"65eea15e30b4559c5844037f2d80bc7356e26e7936cf86a10e09bfd7877fe52c"} Oct 01 09:49:52 crc kubenswrapper[4787]: I1001 09:49:52.549466 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-fqgnk" event={"ID":"7139ac70-d405-449b-924e-01fd94e1375a","Type":"ContainerStarted","Data":"a0e4ddcf51681b54b90b99c198e4e4b175a6b3713abd76fd0ed03871a7ccc2ab"} Oct 01 09:49:52 crc kubenswrapper[4787]: I1001 09:49:52.549496 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-nbxwx" event={"ID":"c186a6f1-61b9-4ab7-8996-5e1d8439a77c","Type":"ContainerStarted","Data":"d20465ec1b75d8032fd4fda359b7acff2b5ebfb62d19e792619337dc3a4e2674"} Oct 01 09:49:52 crc kubenswrapper[4787]: I1001 09:49:52.549508 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-x2v89" event={"ID":"b4902328-fb90-40f9-a84c-5a0359d699d4","Type":"ContainerStarted","Data":"675b50732d2ab7f8b13d1286c99a958e81b4af6855274a7ed6a5d6eef825ff47"} Oct 01 09:49:52 crc kubenswrapper[4787]: I1001 09:49:52.572796 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-75f8d67d86-xjbwt" event={"ID":"20e0b9a6-f51b-4925-bffd-8fa00e9a8ec7","Type":"ContainerStarted","Data":"b296afcb6ffd917b8881463562cf810f808bb0325fa352183891cf7408e2dcee"} Oct 01 09:49:52 crc kubenswrapper[4787]: I1001 09:49:52.588987 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-lbbxd" event={"ID":"b1a0ae18-3c6d-473a-a660-5ff59f6a87bf","Type":"ContainerStarted","Data":"d0f2da03234e2fbc6adbec0f53ae8f3412003a8a31197e1413bc0b98f069a776"} Oct 01 09:49:52 crc kubenswrapper[4787]: I1001 09:49:52.600533 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-qt9bq" event={"ID":"8f52292a-2603-4524-b3fc-8d7b003f95f8","Type":"ContainerStarted","Data":"f939d2062cdc58e54a6923c40546ab56018775ddc745ac3553f6cba8e93272b6"} Oct 01 09:49:52 crc kubenswrapper[4787]: I1001 09:49:52.604251 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-wqs6n" event={"ID":"26b292ae-dcb2-41f6-86b5-2eddfecdf356","Type":"ContainerStarted","Data":"fac9462933d8e9a3ca27b48e7a7770e307a785b3467f08673fa5cad96d6ef666"} Oct 01 09:49:52 crc kubenswrapper[4787]: I1001 09:49:52.612503 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-84c745747f-8qtcw" event={"ID":"57f9bf14-0074-43ba-882d-a9102b047a3b","Type":"ContainerStarted","Data":"20797694defe93b75a34d7463d4cdcb1090d2d26d8a6c5a7233e5c77833f414e"} Oct 01 09:49:52 crc kubenswrapper[4787]: I1001 09:49:52.619357 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-2pxzs" event={"ID":"77e07ae2-1dae-4570-84b8-c137e86ac628","Type":"ContainerStarted","Data":"592f3c59e0a6ca0ecdbe297a98028b7b91dc1290886ab9aff8707a7acd08cedc"} Oct 01 09:49:52 crc kubenswrapper[4787]: I1001 09:49:52.625171 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5f45cd594f-zrtv2" event={"ID":"ae2ba2f0-8422-4a03-b64f-d335f76c5532","Type":"ContainerStarted","Data":"90a94c52a65cdc64a8f5fa42623b129f1e43c776229ff46ceff47a50d68aadeb"} Oct 01 09:49:52 crc kubenswrapper[4787]: E1001 09:49:52.627214 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:ae0bf0aa387864d26d4cbd04d84f6362d53b721348188f096aaa48757419c4f5\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-5f45cd594f-zrtv2" podUID="ae2ba2f0-8422-4a03-b64f-d335f76c5532" Oct 01 09:49:52 crc kubenswrapper[4787]: I1001 09:49:52.641781 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-828zk" event={"ID":"ce71de87-f660-4d9e-845c-1f722a75aa54","Type":"ContainerDied","Data":"eb6899c1145e1df26da8bf8722cbbe6fbfc598e099e31c07ede2d96eb8e80bc3"} Oct 01 09:49:52 crc kubenswrapper[4787]: I1001 09:49:52.641855 4787 scope.go:117] "RemoveContainer" containerID="690a760a08ddfc9c719b917a5782b9472b36455248abaf828731f30bb471b8e4" Oct 01 09:49:52 crc kubenswrapper[4787]: I1001 09:49:52.642043 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-828zk" Oct 01 09:49:52 crc kubenswrapper[4787]: I1001 09:49:52.976159 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-828zk"] Oct 01 09:49:52 crc kubenswrapper[4787]: I1001 09:49:52.991106 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-828zk"] Oct 01 09:49:53 crc kubenswrapper[4787]: I1001 09:49:53.073338 4787 scope.go:117] "RemoveContainer" containerID="283a38eebc39e51af29e296289ac06e4f14a43d346a8fa60307b3521e2d6eab2" Oct 01 09:49:53 crc kubenswrapper[4787]: I1001 09:49:53.158470 4787 scope.go:117] "RemoveContainer" containerID="7104fc1125c4f6296ec1f7983a523d7e97b7bc3be97f62428055208974c5d9f4" Oct 01 09:49:53 crc kubenswrapper[4787]: I1001 09:49:53.680786 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-54fbbfcd44-4jlcz" event={"ID":"8b59a5e1-4f45-4bcd-b9ce-63bae8b5ed57","Type":"ContainerStarted","Data":"ac416f3a5812ddb913c635560e2d317f61fa83c428233d1fd73b8a4bcbfa15a8"} Oct 01 09:49:53 crc kubenswrapper[4787]: I1001 09:49:53.688605 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-84c745747f-8qtcw" event={"ID":"57f9bf14-0074-43ba-882d-a9102b047a3b","Type":"ContainerStarted","Data":"9f8c0701397c4e51680bb27022120f2bdaca39f46e6701c8e89f1ead923c6cc7"} Oct 01 09:49:53 crc kubenswrapper[4787]: I1001 09:49:53.688735 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-84c745747f-8qtcw" Oct 01 09:49:53 crc kubenswrapper[4787]: I1001 09:49:53.693577 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-nbxwx" event={"ID":"c186a6f1-61b9-4ab7-8996-5e1d8439a77c","Type":"ContainerStarted","Data":"f92310772673b743877e8d31caf6dac3ec49b88554cee75c0e9fee3761cc389b"} Oct 01 09:49:53 crc kubenswrapper[4787]: I1001 09:49:53.693744 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-nbxwx" Oct 01 09:49:53 crc kubenswrapper[4787]: I1001 09:49:53.695047 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859cd486d-6h8cm" event={"ID":"24354b04-0daa-4fd4-8d38-b21f64bc2cfe","Type":"ContainerStarted","Data":"8df71b883120bd5d789e6e6ec061fcaead900cbe8d9d268f8903419d9efaa892"} Oct 01 09:49:53 crc kubenswrapper[4787]: I1001 09:49:53.697586 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-qt9bq" event={"ID":"8f52292a-2603-4524-b3fc-8d7b003f95f8","Type":"ContainerStarted","Data":"62c5622d14e173dbe6de532485275fa12f866bc437de47838769f62918866a32"} Oct 01 09:49:53 crc kubenswrapper[4787]: I1001 09:49:53.698124 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-qt9bq" Oct 01 09:49:53 crc kubenswrapper[4787]: I1001 09:49:53.709555 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-wqs6n" event={"ID":"26b292ae-dcb2-41f6-86b5-2eddfecdf356","Type":"ContainerStarted","Data":"516c99640b3fefa33a6e234846be48a6ae3da39213f7297cdf8a871ae9642324"} Oct 01 09:49:53 crc kubenswrapper[4787]: I1001 09:49:53.709633 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-wqs6n" Oct 01 09:49:53 crc kubenswrapper[4787]: I1001 09:49:53.710729 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-84c745747f-8qtcw" podStartSLOduration=4.739168694 podStartE2EDuration="16.710714314s" podCreationTimestamp="2025-10-01 09:49:37 +0000 UTC" firstStartedPulling="2025-10-01 09:49:39.654671639 +0000 UTC m=+811.769815796" lastFinishedPulling="2025-10-01 09:49:51.626217259 +0000 UTC m=+823.741361416" observedRunningTime="2025-10-01 09:49:53.710596461 +0000 UTC m=+825.825740618" watchObservedRunningTime="2025-10-01 09:49:53.710714314 +0000 UTC m=+825.825858471" Oct 01 09:49:53 crc kubenswrapper[4787]: I1001 09:49:53.730556 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-lbbxd" event={"ID":"b1a0ae18-3c6d-473a-a660-5ff59f6a87bf","Type":"ContainerStarted","Data":"c2b7026f96189051d6b6b1f991fdae7a1421eff84f224775198b279a8470dc16"} Oct 01 09:49:53 crc kubenswrapper[4787]: I1001 09:49:53.731326 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-lbbxd" Oct 01 09:49:53 crc kubenswrapper[4787]: I1001 09:49:53.737741 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-2pxzs" event={"ID":"77e07ae2-1dae-4570-84b8-c137e86ac628","Type":"ContainerStarted","Data":"966efa7be945b7828c58a5dd01db6bc93630737e0f904c778fbf92f4c6cd6abe"} Oct 01 09:49:53 crc kubenswrapper[4787]: I1001 09:49:53.738291 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-2pxzs" Oct 01 09:49:53 crc kubenswrapper[4787]: I1001 09:49:53.741984 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-qt9bq" podStartSLOduration=4.310664367 podStartE2EDuration="16.741968077s" podCreationTimestamp="2025-10-01 09:49:37 +0000 UTC" firstStartedPulling="2025-10-01 09:49:39.267429885 +0000 UTC m=+811.382574042" lastFinishedPulling="2025-10-01 09:49:51.698733585 +0000 UTC m=+823.813877752" observedRunningTime="2025-10-01 09:49:53.730946525 +0000 UTC m=+825.846090682" watchObservedRunningTime="2025-10-01 09:49:53.741968077 +0000 UTC m=+825.857112234" Oct 01 09:49:53 crc kubenswrapper[4787]: I1001 09:49:53.752942 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-689b4f76c9-xvrsn" event={"ID":"ed13cb34-e846-41ab-aff7-8e5d4acd2e6b","Type":"ContainerStarted","Data":"ee1aef0134564a60b35e5a83ae5756210f991070239a67eca4c242fe19970670"} Oct 01 09:49:53 crc kubenswrapper[4787]: I1001 09:49:53.753804 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-689b4f76c9-xvrsn" Oct 01 09:49:53 crc kubenswrapper[4787]: I1001 09:49:53.756959 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-nbxwx" podStartSLOduration=4.698538848 podStartE2EDuration="16.756945803s" podCreationTimestamp="2025-10-01 09:49:37 +0000 UTC" firstStartedPulling="2025-10-01 09:49:39.644466546 +0000 UTC m=+811.759610703" lastFinishedPulling="2025-10-01 09:49:51.702873491 +0000 UTC m=+823.818017658" observedRunningTime="2025-10-01 09:49:53.748836984 +0000 UTC m=+825.863981141" watchObservedRunningTime="2025-10-01 09:49:53.756945803 +0000 UTC m=+825.872089960" Oct 01 09:49:53 crc kubenswrapper[4787]: I1001 09:49:53.773659 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-x2v89" event={"ID":"b4902328-fb90-40f9-a84c-5a0359d699d4","Type":"ContainerStarted","Data":"2de212212039681f04f923f4a62f2198cc8672260e71013ebdf998e0a96f6d5e"} Oct 01 09:49:53 crc kubenswrapper[4787]: I1001 09:49:53.773795 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-x2v89" Oct 01 09:49:53 crc kubenswrapper[4787]: I1001 09:49:53.780785 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-wqs6n" podStartSLOduration=4.5886081 podStartE2EDuration="17.780761016s" podCreationTimestamp="2025-10-01 09:49:36 +0000 UTC" firstStartedPulling="2025-10-01 09:49:38.425853381 +0000 UTC m=+810.540997538" lastFinishedPulling="2025-10-01 09:49:51.618006297 +0000 UTC m=+823.733150454" observedRunningTime="2025-10-01 09:49:53.77938033 +0000 UTC m=+825.894524477" watchObservedRunningTime="2025-10-01 09:49:53.780761016 +0000 UTC m=+825.895905173" Oct 01 09:49:53 crc kubenswrapper[4787]: I1001 09:49:53.782036 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-4zm2q" event={"ID":"5024aa9b-39fc-4e58-96fe-c66d054e3d5a","Type":"ContainerStarted","Data":"e4bc8f7f2a08d2f029d97831c8d346f8fa856dc8015ded1b1a2da7c7bd24f7aa"} Oct 01 09:49:53 crc kubenswrapper[4787]: I1001 09:49:53.782169 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-4zm2q" Oct 01 09:49:53 crc kubenswrapper[4787]: I1001 09:49:53.813056 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-5qz2m" event={"ID":"2dfc2636-198f-4b51-b07e-40ff76c1091c","Type":"ContainerStarted","Data":"0a06d3166c6f6f514f2b4fbc818043ddef5e556302551afb4811662e2c664b52"} Oct 01 09:49:53 crc kubenswrapper[4787]: I1001 09:49:53.813123 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-5qz2m" event={"ID":"2dfc2636-198f-4b51-b07e-40ff76c1091c","Type":"ContainerStarted","Data":"bfd3cd420068b2e019ae28d418f0e85b49cb33f171f5e0710c10b554b40d8193"} Oct 01 09:49:53 crc kubenswrapper[4787]: I1001 09:49:53.813138 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-5qz2m" Oct 01 09:49:53 crc kubenswrapper[4787]: I1001 09:49:53.814028 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-lbbxd" podStartSLOduration=4.217101911 podStartE2EDuration="16.814003571s" podCreationTimestamp="2025-10-01 09:49:37 +0000 UTC" firstStartedPulling="2025-10-01 09:49:39.102396569 +0000 UTC m=+811.217540726" lastFinishedPulling="2025-10-01 09:49:51.699298229 +0000 UTC m=+823.814442386" observedRunningTime="2025-10-01 09:49:53.813431196 +0000 UTC m=+825.928575343" watchObservedRunningTime="2025-10-01 09:49:53.814003571 +0000 UTC m=+825.929147728" Oct 01 09:49:53 crc kubenswrapper[4787]: E1001 09:49:53.815863 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:ae0bf0aa387864d26d4cbd04d84f6362d53b721348188f096aaa48757419c4f5\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-5f45cd594f-zrtv2" podUID="ae2ba2f0-8422-4a03-b64f-d335f76c5532" Oct 01 09:49:53 crc kubenswrapper[4787]: I1001 09:49:53.841686 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-x2v89" podStartSLOduration=4.255000297 podStartE2EDuration="16.841640263s" podCreationTimestamp="2025-10-01 09:49:37 +0000 UTC" firstStartedPulling="2025-10-01 09:49:39.102672826 +0000 UTC m=+811.217816983" lastFinishedPulling="2025-10-01 09:49:51.689312792 +0000 UTC m=+823.804456949" observedRunningTime="2025-10-01 09:49:53.837388813 +0000 UTC m=+825.952532970" watchObservedRunningTime="2025-10-01 09:49:53.841640263 +0000 UTC m=+825.956784420" Oct 01 09:49:53 crc kubenswrapper[4787]: I1001 09:49:53.867735 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-689b4f76c9-xvrsn" podStartSLOduration=4.835460601 podStartE2EDuration="16.867687652s" podCreationTimestamp="2025-10-01 09:49:37 +0000 UTC" firstStartedPulling="2025-10-01 09:49:39.658159199 +0000 UTC m=+811.773303356" lastFinishedPulling="2025-10-01 09:49:51.69038625 +0000 UTC m=+823.805530407" observedRunningTime="2025-10-01 09:49:53.865445675 +0000 UTC m=+825.980589832" watchObservedRunningTime="2025-10-01 09:49:53.867687652 +0000 UTC m=+825.982831809" Oct 01 09:49:53 crc kubenswrapper[4787]: I1001 09:49:53.933195 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-2pxzs" podStartSLOduration=4.338632068 podStartE2EDuration="16.933174048s" podCreationTimestamp="2025-10-01 09:49:37 +0000 UTC" firstStartedPulling="2025-10-01 09:49:39.107034248 +0000 UTC m=+811.222178405" lastFinishedPulling="2025-10-01 09:49:51.701576228 +0000 UTC m=+823.816720385" observedRunningTime="2025-10-01 09:49:53.902541329 +0000 UTC m=+826.017685506" watchObservedRunningTime="2025-10-01 09:49:53.933174048 +0000 UTC m=+826.048318205" Oct 01 09:49:53 crc kubenswrapper[4787]: I1001 09:49:53.971837 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-5qz2m" podStartSLOduration=4.939338005 podStartE2EDuration="16.971813832s" podCreationTimestamp="2025-10-01 09:49:37 +0000 UTC" firstStartedPulling="2025-10-01 09:49:39.656764213 +0000 UTC m=+811.771908370" lastFinishedPulling="2025-10-01 09:49:51.68924004 +0000 UTC m=+823.804384197" observedRunningTime="2025-10-01 09:49:53.967211673 +0000 UTC m=+826.082356050" watchObservedRunningTime="2025-10-01 09:49:53.971813832 +0000 UTC m=+826.086957989" Oct 01 09:49:53 crc kubenswrapper[4787]: I1001 09:49:53.997156 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-4zm2q" podStartSLOduration=5.298502746 podStartE2EDuration="16.997137654s" podCreationTimestamp="2025-10-01 09:49:37 +0000 UTC" firstStartedPulling="2025-10-01 09:49:40.002644423 +0000 UTC m=+812.117788590" lastFinishedPulling="2025-10-01 09:49:51.701279341 +0000 UTC m=+823.816423498" observedRunningTime="2025-10-01 09:49:53.996345033 +0000 UTC m=+826.111489190" watchObservedRunningTime="2025-10-01 09:49:53.997137654 +0000 UTC m=+826.112281811" Oct 01 09:49:54 crc kubenswrapper[4787]: I1001 09:49:54.545202 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce71de87-f660-4d9e-845c-1f722a75aa54" path="/var/lib/kubelet/pods/ce71de87-f660-4d9e-845c-1f722a75aa54/volumes" Oct 01 09:49:54 crc kubenswrapper[4787]: I1001 09:49:54.828155 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-t72zp" event={"ID":"9e03d54f-d462-4894-9c3b-8253a861af54","Type":"ContainerStarted","Data":"b4ea269cd9376d76b07b43414458df75ba63781c1c1522bb3268ac1b9b36ba7b"} Oct 01 09:49:54 crc kubenswrapper[4787]: I1001 09:49:54.828317 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-t72zp" Oct 01 09:49:54 crc kubenswrapper[4787]: I1001 09:49:54.832186 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-fqgnk" event={"ID":"7139ac70-d405-449b-924e-01fd94e1375a","Type":"ContainerStarted","Data":"42e85cc58786c37f5e5bec286ecb4933f23c661229f48a3525c62fceb3d841e4"} Oct 01 09:49:54 crc kubenswrapper[4787]: I1001 09:49:54.832660 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-fqgnk" Oct 01 09:49:54 crc kubenswrapper[4787]: I1001 09:49:54.835251 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-689b4f76c9-xvrsn" event={"ID":"ed13cb34-e846-41ab-aff7-8e5d4acd2e6b","Type":"ContainerStarted","Data":"710ba710f72e7a21c26a517006ca520a76840088a08cfa52503d9af28536b698"} Oct 01 09:49:54 crc kubenswrapper[4787]: I1001 09:49:54.837574 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859cd486d-6h8cm" event={"ID":"24354b04-0daa-4fd4-8d38-b21f64bc2cfe","Type":"ContainerStarted","Data":"cd8705882293f29505e04813b14d21aa732bf726c8ec99b416e2d57a91fa5022"} Oct 01 09:49:54 crc kubenswrapper[4787]: I1001 09:49:54.837662 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859cd486d-6h8cm" Oct 01 09:49:54 crc kubenswrapper[4787]: I1001 09:49:54.839370 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-4zm2q" event={"ID":"5024aa9b-39fc-4e58-96fe-c66d054e3d5a","Type":"ContainerStarted","Data":"1613b184b889d5d25b2df6cdfcd7feae4aee8a376e1e4906821a530ff2441452"} Oct 01 09:49:54 crc kubenswrapper[4787]: I1001 09:49:54.843498 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-75f8d67d86-xjbwt" event={"ID":"20e0b9a6-f51b-4925-bffd-8fa00e9a8ec7","Type":"ContainerStarted","Data":"1b4c632130fec7da5c9d32b50d90ab6a7e32bab8c1577829bbb3f28486e6e7d0"} Oct 01 09:49:54 crc kubenswrapper[4787]: I1001 09:49:54.844979 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-75f8d67d86-xjbwt" Oct 01 09:49:54 crc kubenswrapper[4787]: I1001 09:49:54.855933 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-t72zp" podStartSLOduration=5.811021072 podStartE2EDuration="17.85591069s" podCreationTimestamp="2025-10-01 09:49:37 +0000 UTC" firstStartedPulling="2025-10-01 09:49:39.654444583 +0000 UTC m=+811.769588740" lastFinishedPulling="2025-10-01 09:49:51.699334201 +0000 UTC m=+823.814478358" observedRunningTime="2025-10-01 09:49:54.854392061 +0000 UTC m=+826.969536228" watchObservedRunningTime="2025-10-01 09:49:54.85591069 +0000 UTC m=+826.971054847" Oct 01 09:49:54 crc kubenswrapper[4787]: I1001 09:49:54.857307 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-54fbbfcd44-4jlcz" event={"ID":"8b59a5e1-4f45-4bcd-b9ce-63bae8b5ed57","Type":"ContainerStarted","Data":"2531774e2b0ffbdf2a462b759c6d8ddfbdfa91c741463f209e50d4dfe6715559"} Oct 01 09:49:54 crc kubenswrapper[4787]: I1001 09:49:54.858099 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-54fbbfcd44-4jlcz" Oct 01 09:49:54 crc kubenswrapper[4787]: I1001 09:49:54.872117 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-75f8d67d86-xjbwt" podStartSLOduration=5.815330084 podStartE2EDuration="17.872101157s" podCreationTimestamp="2025-10-01 09:49:37 +0000 UTC" firstStartedPulling="2025-10-01 09:49:39.64385082 +0000 UTC m=+811.758994977" lastFinishedPulling="2025-10-01 09:49:51.700621873 +0000 UTC m=+823.815766050" observedRunningTime="2025-10-01 09:49:54.871916681 +0000 UTC m=+826.987060848" watchObservedRunningTime="2025-10-01 09:49:54.872101157 +0000 UTC m=+826.987245314" Oct 01 09:49:54 crc kubenswrapper[4787]: I1001 09:49:54.894913 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859cd486d-6h8cm" podStartSLOduration=4.795586216 podStartE2EDuration="17.894881193s" podCreationTimestamp="2025-10-01 09:49:37 +0000 UTC" firstStartedPulling="2025-10-01 09:49:38.60224687 +0000 UTC m=+810.717391027" lastFinishedPulling="2025-10-01 09:49:51.701541847 +0000 UTC m=+823.816686004" observedRunningTime="2025-10-01 09:49:54.890025137 +0000 UTC m=+827.005169294" watchObservedRunningTime="2025-10-01 09:49:54.894881193 +0000 UTC m=+827.010025350" Oct 01 09:49:54 crc kubenswrapper[4787]: I1001 09:49:54.916126 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-fqgnk" podStartSLOduration=6.168156201 podStartE2EDuration="17.916100339s" podCreationTimestamp="2025-10-01 09:49:37 +0000 UTC" firstStartedPulling="2025-10-01 09:49:40.01033767 +0000 UTC m=+812.125481827" lastFinishedPulling="2025-10-01 09:49:51.758281788 +0000 UTC m=+823.873425965" observedRunningTime="2025-10-01 09:49:54.908438351 +0000 UTC m=+827.023582538" watchObservedRunningTime="2025-10-01 09:49:54.916100339 +0000 UTC m=+827.031244496" Oct 01 09:49:54 crc kubenswrapper[4787]: I1001 09:49:54.925851 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-54fbbfcd44-4jlcz" podStartSLOduration=5.409641715 podStartE2EDuration="17.925831849s" podCreationTimestamp="2025-10-01 09:49:37 +0000 UTC" firstStartedPulling="2025-10-01 09:49:39.101850464 +0000 UTC m=+811.216994621" lastFinishedPulling="2025-10-01 09:49:51.618040598 +0000 UTC m=+823.733184755" observedRunningTime="2025-10-01 09:49:54.923492608 +0000 UTC m=+827.038636765" watchObservedRunningTime="2025-10-01 09:49:54.925831849 +0000 UTC m=+827.040976006" Oct 01 09:49:55 crc kubenswrapper[4787]: I1001 09:49:55.870614 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-9tdp2" event={"ID":"88d93cbc-27c2-47be-a882-6607925d4391","Type":"ContainerStarted","Data":"ceaf099dca71b2b30dab8a6f997f975b6e18c9c41db2545447ba7f4338d42038"} Oct 01 09:49:56 crc kubenswrapper[4787]: I1001 09:49:56.882889 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-qmr4v" event={"ID":"b54a691a-cd96-4ae4-a9ce-65662f58cd60","Type":"ContainerStarted","Data":"78f157e5aebc0b12b72eddbca16729f09fe6aa605c0e31affe323ef4d84e01bb"} Oct 01 09:49:56 crc kubenswrapper[4787]: I1001 09:49:56.928829 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-9tdp2" podStartSLOduration=4.529607341 podStartE2EDuration="19.928791545s" podCreationTimestamp="2025-10-01 09:49:37 +0000 UTC" firstStartedPulling="2025-10-01 09:49:39.682056854 +0000 UTC m=+811.797201011" lastFinishedPulling="2025-10-01 09:49:55.081241048 +0000 UTC m=+827.196385215" observedRunningTime="2025-10-01 09:49:55.891810404 +0000 UTC m=+828.006954571" watchObservedRunningTime="2025-10-01 09:49:56.928791545 +0000 UTC m=+829.043935722" Oct 01 09:49:56 crc kubenswrapper[4787]: I1001 09:49:56.934038 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-qmr4v" podStartSLOduration=3.392469222 podStartE2EDuration="19.9340217s" podCreationTimestamp="2025-10-01 09:49:37 +0000 UTC" firstStartedPulling="2025-10-01 09:49:39.665804785 +0000 UTC m=+811.780948942" lastFinishedPulling="2025-10-01 09:49:56.207357263 +0000 UTC m=+828.322501420" observedRunningTime="2025-10-01 09:49:56.925301856 +0000 UTC m=+829.040446063" watchObservedRunningTime="2025-10-01 09:49:56.9340217 +0000 UTC m=+829.049165867" Oct 01 09:49:57 crc kubenswrapper[4787]: I1001 09:49:57.365158 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859cd486d-6h8cm" Oct 01 09:49:57 crc kubenswrapper[4787]: I1001 09:49:57.391411 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-wqs6n" Oct 01 09:49:57 crc kubenswrapper[4787]: I1001 09:49:57.442042 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-2pxzs" Oct 01 09:49:57 crc kubenswrapper[4787]: I1001 09:49:57.479280 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-x2v89" Oct 01 09:49:57 crc kubenswrapper[4787]: I1001 09:49:57.517986 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-lbbxd" Oct 01 09:49:57 crc kubenswrapper[4787]: I1001 09:49:57.540698 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-qt9bq" Oct 01 09:49:57 crc kubenswrapper[4787]: I1001 09:49:57.812454 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-qmr4v" Oct 01 09:49:57 crc kubenswrapper[4787]: I1001 09:49:57.815016 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-54fbbfcd44-4jlcz" Oct 01 09:49:57 crc kubenswrapper[4787]: I1001 09:49:57.961113 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-nbxwx" Oct 01 09:49:58 crc kubenswrapper[4787]: I1001 09:49:58.032387 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-9tdp2" Oct 01 09:49:58 crc kubenswrapper[4787]: I1001 09:49:58.042101 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-75f8d67d86-xjbwt" Oct 01 09:49:58 crc kubenswrapper[4787]: I1001 09:49:58.066053 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-5qz2m" Oct 01 09:49:58 crc kubenswrapper[4787]: I1001 09:49:58.204378 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-t72zp" Oct 01 09:49:58 crc kubenswrapper[4787]: I1001 09:49:58.314182 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-84c745747f-8qtcw" Oct 01 09:49:58 crc kubenswrapper[4787]: I1001 09:49:58.380130 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-689b4f76c9-xvrsn" Oct 01 09:49:58 crc kubenswrapper[4787]: I1001 09:49:58.449318 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-4zm2q" Oct 01 09:49:58 crc kubenswrapper[4787]: I1001 09:49:58.515951 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-fqgnk" Oct 01 09:49:58 crc kubenswrapper[4787]: I1001 09:49:58.900318 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-659bb84579xswrt" event={"ID":"2d5ffd79-67fa-4dd0-81a7-b44156946b86","Type":"ContainerStarted","Data":"2e0c838f5ca7d618ef1954b4cb4311f6da4379f46bcee8d214fff593e79335fa"} Oct 01 09:49:58 crc kubenswrapper[4787]: I1001 09:49:58.900587 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-659bb84579xswrt" Oct 01 09:49:58 crc kubenswrapper[4787]: I1001 09:49:58.902595 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xvd4r" event={"ID":"f31cb2c2-eabc-4bca-88c5-2aa194fb13b5","Type":"ContainerStarted","Data":"35a0a73fc0c7a519c920f69e91d109228c433b10e9556ccdffac684850433afc"} Oct 01 09:49:58 crc kubenswrapper[4787]: I1001 09:49:58.937357 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-659bb84579xswrt" podStartSLOduration=3.780078917 podStartE2EDuration="21.937330926s" podCreationTimestamp="2025-10-01 09:49:37 +0000 UTC" firstStartedPulling="2025-10-01 09:49:40.027862632 +0000 UTC m=+812.143006789" lastFinishedPulling="2025-10-01 09:49:58.185114641 +0000 UTC m=+830.300258798" observedRunningTime="2025-10-01 09:49:58.932211164 +0000 UTC m=+831.047355331" watchObservedRunningTime="2025-10-01 09:49:58.937330926 +0000 UTC m=+831.052475083" Oct 01 09:49:58 crc kubenswrapper[4787]: I1001 09:49:58.950940 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xvd4r" podStartSLOduration=3.797229708 podStartE2EDuration="21.950917466s" podCreationTimestamp="2025-10-01 09:49:37 +0000 UTC" firstStartedPulling="2025-10-01 09:49:40.010440583 +0000 UTC m=+812.125584740" lastFinishedPulling="2025-10-01 09:49:58.164128341 +0000 UTC m=+830.279272498" observedRunningTime="2025-10-01 09:49:58.948902423 +0000 UTC m=+831.064046600" watchObservedRunningTime="2025-10-01 09:49:58.950917466 +0000 UTC m=+831.066061623" Oct 01 09:50:00 crc kubenswrapper[4787]: I1001 09:50:00.919191 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-z2gzf" event={"ID":"a74d851b-a805-4138-9bf4-1a4546a2f30a","Type":"ContainerStarted","Data":"7cb3277e77381c1233610350e3b006dd9ded328a8227d0c91562932c5e066c5d"} Oct 01 09:50:00 crc kubenswrapper[4787]: I1001 09:50:00.920225 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-z2gzf" Oct 01 09:50:00 crc kubenswrapper[4787]: I1001 09:50:00.922145 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-4mgdw" event={"ID":"fb776e43-5f95-44f0-96c0-945b2681d8c1","Type":"ContainerStarted","Data":"f804baa943f7b5e44d754dc7323ef0ef2a5e422afc0b5577e078904e22079777"} Oct 01 09:50:00 crc kubenswrapper[4787]: I1001 09:50:00.922333 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-4mgdw" Oct 01 09:50:00 crc kubenswrapper[4787]: I1001 09:50:00.940119 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-z2gzf" podStartSLOduration=4.022141434 podStartE2EDuration="23.940102477s" podCreationTimestamp="2025-10-01 09:49:37 +0000 UTC" firstStartedPulling="2025-10-01 09:49:40.027516092 +0000 UTC m=+812.142660249" lastFinishedPulling="2025-10-01 09:49:59.945477135 +0000 UTC m=+832.060621292" observedRunningTime="2025-10-01 09:50:00.934334139 +0000 UTC m=+833.049478306" watchObservedRunningTime="2025-10-01 09:50:00.940102477 +0000 UTC m=+833.055246634" Oct 01 09:50:00 crc kubenswrapper[4787]: I1001 09:50:00.951815 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-4mgdw" podStartSLOduration=3.669678915 podStartE2EDuration="23.951795578s" podCreationTimestamp="2025-10-01 09:49:37 +0000 UTC" firstStartedPulling="2025-10-01 09:49:39.668737141 +0000 UTC m=+811.783881298" lastFinishedPulling="2025-10-01 09:49:59.950853814 +0000 UTC m=+832.065997961" observedRunningTime="2025-10-01 09:50:00.947305463 +0000 UTC m=+833.062449640" watchObservedRunningTime="2025-10-01 09:50:00.951795578 +0000 UTC m=+833.066939735" Oct 01 09:50:05 crc kubenswrapper[4787]: I1001 09:50:05.954393 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5f45cd594f-zrtv2" event={"ID":"ae2ba2f0-8422-4a03-b64f-d335f76c5532","Type":"ContainerStarted","Data":"1979e7efd6a9a507da3f01d79de474a6a4293cf6515054775e43f88749bdf8bc"} Oct 01 09:50:05 crc kubenswrapper[4787]: I1001 09:50:05.955437 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5f45cd594f-zrtv2" Oct 01 09:50:05 crc kubenswrapper[4787]: I1001 09:50:05.975276 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-5f45cd594f-zrtv2" podStartSLOduration=3.131305162 podStartE2EDuration="28.975257933s" podCreationTimestamp="2025-10-01 09:49:37 +0000 UTC" firstStartedPulling="2025-10-01 09:49:39.257475028 +0000 UTC m=+811.372619185" lastFinishedPulling="2025-10-01 09:50:05.101427799 +0000 UTC m=+837.216571956" observedRunningTime="2025-10-01 09:50:05.970986934 +0000 UTC m=+838.086131091" watchObservedRunningTime="2025-10-01 09:50:05.975257933 +0000 UTC m=+838.090402090" Oct 01 09:50:07 crc kubenswrapper[4787]: I1001 09:50:07.814509 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-qmr4v" Oct 01 09:50:08 crc kubenswrapper[4787]: I1001 09:50:08.035865 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-9tdp2" Oct 01 09:50:08 crc kubenswrapper[4787]: I1001 09:50:08.504978 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-z2gzf" Oct 01 09:50:08 crc kubenswrapper[4787]: I1001 09:50:08.538586 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-659bb84579xswrt" Oct 01 09:50:08 crc kubenswrapper[4787]: I1001 09:50:08.635715 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-4mgdw" Oct 01 09:50:17 crc kubenswrapper[4787]: I1001 09:50:17.632684 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-5f45cd594f-zrtv2" Oct 01 09:50:33 crc kubenswrapper[4787]: I1001 09:50:33.432454 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8b69cf79-6w6rh"] Oct 01 09:50:33 crc kubenswrapper[4787]: E1001 09:50:33.435815 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce71de87-f660-4d9e-845c-1f722a75aa54" containerName="registry-server" Oct 01 09:50:33 crc kubenswrapper[4787]: I1001 09:50:33.435940 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce71de87-f660-4d9e-845c-1f722a75aa54" containerName="registry-server" Oct 01 09:50:33 crc kubenswrapper[4787]: E1001 09:50:33.436108 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce71de87-f660-4d9e-845c-1f722a75aa54" containerName="extract-utilities" Oct 01 09:50:33 crc kubenswrapper[4787]: I1001 09:50:33.436210 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce71de87-f660-4d9e-845c-1f722a75aa54" containerName="extract-utilities" Oct 01 09:50:33 crc kubenswrapper[4787]: E1001 09:50:33.436317 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce71de87-f660-4d9e-845c-1f722a75aa54" containerName="extract-content" Oct 01 09:50:33 crc kubenswrapper[4787]: I1001 09:50:33.436399 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce71de87-f660-4d9e-845c-1f722a75aa54" containerName="extract-content" Oct 01 09:50:33 crc kubenswrapper[4787]: I1001 09:50:33.436652 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce71de87-f660-4d9e-845c-1f722a75aa54" containerName="registry-server" Oct 01 09:50:33 crc kubenswrapper[4787]: I1001 09:50:33.437677 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8b69cf79-6w6rh" Oct 01 09:50:33 crc kubenswrapper[4787]: I1001 09:50:33.442541 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 01 09:50:33 crc kubenswrapper[4787]: I1001 09:50:33.442615 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-rg7zw" Oct 01 09:50:33 crc kubenswrapper[4787]: I1001 09:50:33.442743 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 01 09:50:33 crc kubenswrapper[4787]: I1001 09:50:33.443019 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 01 09:50:33 crc kubenswrapper[4787]: I1001 09:50:33.450490 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8b69cf79-6w6rh"] Oct 01 09:50:33 crc kubenswrapper[4787]: I1001 09:50:33.487609 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dr9d\" (UniqueName: \"kubernetes.io/projected/0bdbc459-b8b9-4a6f-9462-61bd54c87b1d-kube-api-access-8dr9d\") pod \"dnsmasq-dns-b8b69cf79-6w6rh\" (UID: \"0bdbc459-b8b9-4a6f-9462-61bd54c87b1d\") " pod="openstack/dnsmasq-dns-b8b69cf79-6w6rh" Oct 01 09:50:33 crc kubenswrapper[4787]: I1001 09:50:33.488259 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bdbc459-b8b9-4a6f-9462-61bd54c87b1d-config\") pod \"dnsmasq-dns-b8b69cf79-6w6rh\" (UID: \"0bdbc459-b8b9-4a6f-9462-61bd54c87b1d\") " pod="openstack/dnsmasq-dns-b8b69cf79-6w6rh" Oct 01 09:50:33 crc kubenswrapper[4787]: I1001 09:50:33.570910 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d5f6f49c7-jqk5n"] Oct 01 09:50:33 crc kubenswrapper[4787]: I1001 09:50:33.572538 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d5f6f49c7-jqk5n" Oct 01 09:50:33 crc kubenswrapper[4787]: I1001 09:50:33.576976 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 01 09:50:33 crc kubenswrapper[4787]: I1001 09:50:33.590032 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d5f6f49c7-jqk5n"] Oct 01 09:50:33 crc kubenswrapper[4787]: I1001 09:50:33.590946 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dr9d\" (UniqueName: \"kubernetes.io/projected/0bdbc459-b8b9-4a6f-9462-61bd54c87b1d-kube-api-access-8dr9d\") pod \"dnsmasq-dns-b8b69cf79-6w6rh\" (UID: \"0bdbc459-b8b9-4a6f-9462-61bd54c87b1d\") " pod="openstack/dnsmasq-dns-b8b69cf79-6w6rh" Oct 01 09:50:33 crc kubenswrapper[4787]: I1001 09:50:33.590988 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bdbc459-b8b9-4a6f-9462-61bd54c87b1d-config\") pod \"dnsmasq-dns-b8b69cf79-6w6rh\" (UID: \"0bdbc459-b8b9-4a6f-9462-61bd54c87b1d\") " pod="openstack/dnsmasq-dns-b8b69cf79-6w6rh" Oct 01 09:50:33 crc kubenswrapper[4787]: I1001 09:50:33.592262 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bdbc459-b8b9-4a6f-9462-61bd54c87b1d-config\") pod \"dnsmasq-dns-b8b69cf79-6w6rh\" (UID: \"0bdbc459-b8b9-4a6f-9462-61bd54c87b1d\") " pod="openstack/dnsmasq-dns-b8b69cf79-6w6rh" Oct 01 09:50:33 crc kubenswrapper[4787]: I1001 09:50:33.647930 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dr9d\" (UniqueName: \"kubernetes.io/projected/0bdbc459-b8b9-4a6f-9462-61bd54c87b1d-kube-api-access-8dr9d\") pod \"dnsmasq-dns-b8b69cf79-6w6rh\" (UID: \"0bdbc459-b8b9-4a6f-9462-61bd54c87b1d\") " pod="openstack/dnsmasq-dns-b8b69cf79-6w6rh" Oct 01 09:50:33 crc kubenswrapper[4787]: I1001 09:50:33.693061 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lh64v\" (UniqueName: \"kubernetes.io/projected/47914662-d598-4ccf-af68-6d2d3a2d75b3-kube-api-access-lh64v\") pod \"dnsmasq-dns-d5f6f49c7-jqk5n\" (UID: \"47914662-d598-4ccf-af68-6d2d3a2d75b3\") " pod="openstack/dnsmasq-dns-d5f6f49c7-jqk5n" Oct 01 09:50:33 crc kubenswrapper[4787]: I1001 09:50:33.693151 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47914662-d598-4ccf-af68-6d2d3a2d75b3-config\") pod \"dnsmasq-dns-d5f6f49c7-jqk5n\" (UID: \"47914662-d598-4ccf-af68-6d2d3a2d75b3\") " pod="openstack/dnsmasq-dns-d5f6f49c7-jqk5n" Oct 01 09:50:33 crc kubenswrapper[4787]: I1001 09:50:33.693181 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/47914662-d598-4ccf-af68-6d2d3a2d75b3-dns-svc\") pod \"dnsmasq-dns-d5f6f49c7-jqk5n\" (UID: \"47914662-d598-4ccf-af68-6d2d3a2d75b3\") " pod="openstack/dnsmasq-dns-d5f6f49c7-jqk5n" Oct 01 09:50:33 crc kubenswrapper[4787]: I1001 09:50:33.794399 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lh64v\" (UniqueName: \"kubernetes.io/projected/47914662-d598-4ccf-af68-6d2d3a2d75b3-kube-api-access-lh64v\") pod \"dnsmasq-dns-d5f6f49c7-jqk5n\" (UID: \"47914662-d598-4ccf-af68-6d2d3a2d75b3\") " pod="openstack/dnsmasq-dns-d5f6f49c7-jqk5n" Oct 01 09:50:33 crc kubenswrapper[4787]: I1001 09:50:33.794468 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47914662-d598-4ccf-af68-6d2d3a2d75b3-config\") pod \"dnsmasq-dns-d5f6f49c7-jqk5n\" (UID: \"47914662-d598-4ccf-af68-6d2d3a2d75b3\") " pod="openstack/dnsmasq-dns-d5f6f49c7-jqk5n" Oct 01 09:50:33 crc kubenswrapper[4787]: I1001 09:50:33.794492 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/47914662-d598-4ccf-af68-6d2d3a2d75b3-dns-svc\") pod \"dnsmasq-dns-d5f6f49c7-jqk5n\" (UID: \"47914662-d598-4ccf-af68-6d2d3a2d75b3\") " pod="openstack/dnsmasq-dns-d5f6f49c7-jqk5n" Oct 01 09:50:33 crc kubenswrapper[4787]: I1001 09:50:33.795405 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/47914662-d598-4ccf-af68-6d2d3a2d75b3-dns-svc\") pod \"dnsmasq-dns-d5f6f49c7-jqk5n\" (UID: \"47914662-d598-4ccf-af68-6d2d3a2d75b3\") " pod="openstack/dnsmasq-dns-d5f6f49c7-jqk5n" Oct 01 09:50:33 crc kubenswrapper[4787]: I1001 09:50:33.795428 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47914662-d598-4ccf-af68-6d2d3a2d75b3-config\") pod \"dnsmasq-dns-d5f6f49c7-jqk5n\" (UID: \"47914662-d598-4ccf-af68-6d2d3a2d75b3\") " pod="openstack/dnsmasq-dns-d5f6f49c7-jqk5n" Oct 01 09:50:33 crc kubenswrapper[4787]: I1001 09:50:33.811777 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lh64v\" (UniqueName: \"kubernetes.io/projected/47914662-d598-4ccf-af68-6d2d3a2d75b3-kube-api-access-lh64v\") pod \"dnsmasq-dns-d5f6f49c7-jqk5n\" (UID: \"47914662-d598-4ccf-af68-6d2d3a2d75b3\") " pod="openstack/dnsmasq-dns-d5f6f49c7-jqk5n" Oct 01 09:50:33 crc kubenswrapper[4787]: I1001 09:50:33.816095 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8b69cf79-6w6rh" Oct 01 09:50:33 crc kubenswrapper[4787]: I1001 09:50:33.973767 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d5f6f49c7-jqk5n" Oct 01 09:50:34 crc kubenswrapper[4787]: I1001 09:50:34.261700 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8b69cf79-6w6rh"] Oct 01 09:50:34 crc kubenswrapper[4787]: I1001 09:50:34.389450 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d5f6f49c7-jqk5n"] Oct 01 09:50:34 crc kubenswrapper[4787]: W1001 09:50:34.399840 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod47914662_d598_4ccf_af68_6d2d3a2d75b3.slice/crio-e1cecaaac1a7d6793e33a1bfe9714746666ee7d8e097b27a8b5dc79a7c37c358 WatchSource:0}: Error finding container e1cecaaac1a7d6793e33a1bfe9714746666ee7d8e097b27a8b5dc79a7c37c358: Status 404 returned error can't find the container with id e1cecaaac1a7d6793e33a1bfe9714746666ee7d8e097b27a8b5dc79a7c37c358 Oct 01 09:50:35 crc kubenswrapper[4787]: I1001 09:50:35.178189 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8b69cf79-6w6rh" event={"ID":"0bdbc459-b8b9-4a6f-9462-61bd54c87b1d","Type":"ContainerStarted","Data":"9004515597def8ef11563c7909dedc26484b860ac226358d3eb15cb9c70bdf6d"} Oct 01 09:50:35 crc kubenswrapper[4787]: I1001 09:50:35.182177 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d5f6f49c7-jqk5n" event={"ID":"47914662-d598-4ccf-af68-6d2d3a2d75b3","Type":"ContainerStarted","Data":"e1cecaaac1a7d6793e33a1bfe9714746666ee7d8e097b27a8b5dc79a7c37c358"} Oct 01 09:50:36 crc kubenswrapper[4787]: I1001 09:50:36.356990 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8b69cf79-6w6rh"] Oct 01 09:50:36 crc kubenswrapper[4787]: I1001 09:50:36.388365 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b6f94bdfc-jdsn9"] Oct 01 09:50:36 crc kubenswrapper[4787]: I1001 09:50:36.389498 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b6f94bdfc-jdsn9" Oct 01 09:50:36 crc kubenswrapper[4787]: I1001 09:50:36.408128 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b6f94bdfc-jdsn9"] Oct 01 09:50:36 crc kubenswrapper[4787]: I1001 09:50:36.457095 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a5c38d07-c1a7-46fd-bcce-1680eff35b1f-dns-svc\") pod \"dnsmasq-dns-b6f94bdfc-jdsn9\" (UID: \"a5c38d07-c1a7-46fd-bcce-1680eff35b1f\") " pod="openstack/dnsmasq-dns-b6f94bdfc-jdsn9" Oct 01 09:50:36 crc kubenswrapper[4787]: I1001 09:50:36.457407 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdmbr\" (UniqueName: \"kubernetes.io/projected/a5c38d07-c1a7-46fd-bcce-1680eff35b1f-kube-api-access-fdmbr\") pod \"dnsmasq-dns-b6f94bdfc-jdsn9\" (UID: \"a5c38d07-c1a7-46fd-bcce-1680eff35b1f\") " pod="openstack/dnsmasq-dns-b6f94bdfc-jdsn9" Oct 01 09:50:36 crc kubenswrapper[4787]: I1001 09:50:36.457503 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5c38d07-c1a7-46fd-bcce-1680eff35b1f-config\") pod \"dnsmasq-dns-b6f94bdfc-jdsn9\" (UID: \"a5c38d07-c1a7-46fd-bcce-1680eff35b1f\") " pod="openstack/dnsmasq-dns-b6f94bdfc-jdsn9" Oct 01 09:50:36 crc kubenswrapper[4787]: I1001 09:50:36.561436 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a5c38d07-c1a7-46fd-bcce-1680eff35b1f-dns-svc\") pod \"dnsmasq-dns-b6f94bdfc-jdsn9\" (UID: \"a5c38d07-c1a7-46fd-bcce-1680eff35b1f\") " pod="openstack/dnsmasq-dns-b6f94bdfc-jdsn9" Oct 01 09:50:36 crc kubenswrapper[4787]: I1001 09:50:36.562855 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdmbr\" (UniqueName: \"kubernetes.io/projected/a5c38d07-c1a7-46fd-bcce-1680eff35b1f-kube-api-access-fdmbr\") pod \"dnsmasq-dns-b6f94bdfc-jdsn9\" (UID: \"a5c38d07-c1a7-46fd-bcce-1680eff35b1f\") " pod="openstack/dnsmasq-dns-b6f94bdfc-jdsn9" Oct 01 09:50:36 crc kubenswrapper[4787]: I1001 09:50:36.562934 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5c38d07-c1a7-46fd-bcce-1680eff35b1f-config\") pod \"dnsmasq-dns-b6f94bdfc-jdsn9\" (UID: \"a5c38d07-c1a7-46fd-bcce-1680eff35b1f\") " pod="openstack/dnsmasq-dns-b6f94bdfc-jdsn9" Oct 01 09:50:36 crc kubenswrapper[4787]: I1001 09:50:36.563565 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a5c38d07-c1a7-46fd-bcce-1680eff35b1f-dns-svc\") pod \"dnsmasq-dns-b6f94bdfc-jdsn9\" (UID: \"a5c38d07-c1a7-46fd-bcce-1680eff35b1f\") " pod="openstack/dnsmasq-dns-b6f94bdfc-jdsn9" Oct 01 09:50:36 crc kubenswrapper[4787]: I1001 09:50:36.568635 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5c38d07-c1a7-46fd-bcce-1680eff35b1f-config\") pod \"dnsmasq-dns-b6f94bdfc-jdsn9\" (UID: \"a5c38d07-c1a7-46fd-bcce-1680eff35b1f\") " pod="openstack/dnsmasq-dns-b6f94bdfc-jdsn9" Oct 01 09:50:36 crc kubenswrapper[4787]: I1001 09:50:36.618736 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdmbr\" (UniqueName: \"kubernetes.io/projected/a5c38d07-c1a7-46fd-bcce-1680eff35b1f-kube-api-access-fdmbr\") pod \"dnsmasq-dns-b6f94bdfc-jdsn9\" (UID: \"a5c38d07-c1a7-46fd-bcce-1680eff35b1f\") " pod="openstack/dnsmasq-dns-b6f94bdfc-jdsn9" Oct 01 09:50:36 crc kubenswrapper[4787]: I1001 09:50:36.688769 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d5f6f49c7-jqk5n"] Oct 01 09:50:36 crc kubenswrapper[4787]: I1001 09:50:36.717432 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77795d58f5-kh2gz"] Oct 01 09:50:36 crc kubenswrapper[4787]: I1001 09:50:36.717777 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b6f94bdfc-jdsn9" Oct 01 09:50:36 crc kubenswrapper[4787]: I1001 09:50:36.719413 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77795d58f5-kh2gz" Oct 01 09:50:36 crc kubenswrapper[4787]: I1001 09:50:36.732950 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77795d58f5-kh2gz"] Oct 01 09:50:36 crc kubenswrapper[4787]: I1001 09:50:36.871889 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/105b8088-a178-492b-aefd-20a5e6ad0dc8-dns-svc\") pod \"dnsmasq-dns-77795d58f5-kh2gz\" (UID: \"105b8088-a178-492b-aefd-20a5e6ad0dc8\") " pod="openstack/dnsmasq-dns-77795d58f5-kh2gz" Oct 01 09:50:36 crc kubenswrapper[4787]: I1001 09:50:36.871961 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/105b8088-a178-492b-aefd-20a5e6ad0dc8-config\") pod \"dnsmasq-dns-77795d58f5-kh2gz\" (UID: \"105b8088-a178-492b-aefd-20a5e6ad0dc8\") " pod="openstack/dnsmasq-dns-77795d58f5-kh2gz" Oct 01 09:50:36 crc kubenswrapper[4787]: I1001 09:50:36.871996 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnjkn\" (UniqueName: \"kubernetes.io/projected/105b8088-a178-492b-aefd-20a5e6ad0dc8-kube-api-access-rnjkn\") pod \"dnsmasq-dns-77795d58f5-kh2gz\" (UID: \"105b8088-a178-492b-aefd-20a5e6ad0dc8\") " pod="openstack/dnsmasq-dns-77795d58f5-kh2gz" Oct 01 09:50:36 crc kubenswrapper[4787]: I1001 09:50:36.973989 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/105b8088-a178-492b-aefd-20a5e6ad0dc8-dns-svc\") pod \"dnsmasq-dns-77795d58f5-kh2gz\" (UID: \"105b8088-a178-492b-aefd-20a5e6ad0dc8\") " pod="openstack/dnsmasq-dns-77795d58f5-kh2gz" Oct 01 09:50:36 crc kubenswrapper[4787]: I1001 09:50:36.974068 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/105b8088-a178-492b-aefd-20a5e6ad0dc8-config\") pod \"dnsmasq-dns-77795d58f5-kh2gz\" (UID: \"105b8088-a178-492b-aefd-20a5e6ad0dc8\") " pod="openstack/dnsmasq-dns-77795d58f5-kh2gz" Oct 01 09:50:36 crc kubenswrapper[4787]: I1001 09:50:36.974310 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnjkn\" (UniqueName: \"kubernetes.io/projected/105b8088-a178-492b-aefd-20a5e6ad0dc8-kube-api-access-rnjkn\") pod \"dnsmasq-dns-77795d58f5-kh2gz\" (UID: \"105b8088-a178-492b-aefd-20a5e6ad0dc8\") " pod="openstack/dnsmasq-dns-77795d58f5-kh2gz" Oct 01 09:50:36 crc kubenswrapper[4787]: I1001 09:50:36.983895 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/105b8088-a178-492b-aefd-20a5e6ad0dc8-config\") pod \"dnsmasq-dns-77795d58f5-kh2gz\" (UID: \"105b8088-a178-492b-aefd-20a5e6ad0dc8\") " pod="openstack/dnsmasq-dns-77795d58f5-kh2gz" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.141456 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/105b8088-a178-492b-aefd-20a5e6ad0dc8-dns-svc\") pod \"dnsmasq-dns-77795d58f5-kh2gz\" (UID: \"105b8088-a178-492b-aefd-20a5e6ad0dc8\") " pod="openstack/dnsmasq-dns-77795d58f5-kh2gz" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.165881 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnjkn\" (UniqueName: \"kubernetes.io/projected/105b8088-a178-492b-aefd-20a5e6ad0dc8-kube-api-access-rnjkn\") pod \"dnsmasq-dns-77795d58f5-kh2gz\" (UID: \"105b8088-a178-492b-aefd-20a5e6ad0dc8\") " pod="openstack/dnsmasq-dns-77795d58f5-kh2gz" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.351790 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77795d58f5-kh2gz" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.360216 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b6f94bdfc-jdsn9"] Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.549405 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.554170 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.556790 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.557600 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-zsk5j" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.557820 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.558006 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.558426 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.558497 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.558636 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.558807 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.687126 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e762572f-d5b9-462e-9953-7143f648c9ae-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " pod="openstack/rabbitmq-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.687175 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e762572f-d5b9-462e-9953-7143f648c9ae-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " pod="openstack/rabbitmq-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.687232 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e762572f-d5b9-462e-9953-7143f648c9ae-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " pod="openstack/rabbitmq-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.687259 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e762572f-d5b9-462e-9953-7143f648c9ae-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " pod="openstack/rabbitmq-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.687281 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e762572f-d5b9-462e-9953-7143f648c9ae-config-data\") pod \"rabbitmq-server-0\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " pod="openstack/rabbitmq-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.687314 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e762572f-d5b9-462e-9953-7143f648c9ae-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " pod="openstack/rabbitmq-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.687333 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e762572f-d5b9-462e-9953-7143f648c9ae-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " pod="openstack/rabbitmq-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.687347 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6qkc\" (UniqueName: \"kubernetes.io/projected/e762572f-d5b9-462e-9953-7143f648c9ae-kube-api-access-c6qkc\") pod \"rabbitmq-server-0\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " pod="openstack/rabbitmq-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.687366 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " pod="openstack/rabbitmq-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.687387 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e762572f-d5b9-462e-9953-7143f648c9ae-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " pod="openstack/rabbitmq-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.687411 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e762572f-d5b9-462e-9953-7143f648c9ae-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " pod="openstack/rabbitmq-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.788600 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e762572f-d5b9-462e-9953-7143f648c9ae-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " pod="openstack/rabbitmq-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.788660 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e762572f-d5b9-462e-9953-7143f648c9ae-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " pod="openstack/rabbitmq-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.788691 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e762572f-d5b9-462e-9953-7143f648c9ae-config-data\") pod \"rabbitmq-server-0\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " pod="openstack/rabbitmq-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.788738 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e762572f-d5b9-462e-9953-7143f648c9ae-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " pod="openstack/rabbitmq-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.788776 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e762572f-d5b9-462e-9953-7143f648c9ae-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " pod="openstack/rabbitmq-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.788796 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6qkc\" (UniqueName: \"kubernetes.io/projected/e762572f-d5b9-462e-9953-7143f648c9ae-kube-api-access-c6qkc\") pod \"rabbitmq-server-0\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " pod="openstack/rabbitmq-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.788822 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " pod="openstack/rabbitmq-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.789708 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e762572f-d5b9-462e-9953-7143f648c9ae-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " pod="openstack/rabbitmq-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.790027 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.790240 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e762572f-d5b9-462e-9953-7143f648c9ae-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " pod="openstack/rabbitmq-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.790313 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e762572f-d5b9-462e-9953-7143f648c9ae-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " pod="openstack/rabbitmq-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.790426 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e762572f-d5b9-462e-9953-7143f648c9ae-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " pod="openstack/rabbitmq-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.790471 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e762572f-d5b9-462e-9953-7143f648c9ae-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " pod="openstack/rabbitmq-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.790989 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e762572f-d5b9-462e-9953-7143f648c9ae-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " pod="openstack/rabbitmq-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.791452 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e762572f-d5b9-462e-9953-7143f648c9ae-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " pod="openstack/rabbitmq-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.791479 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e762572f-d5b9-462e-9953-7143f648c9ae-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " pod="openstack/rabbitmq-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.793043 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e762572f-d5b9-462e-9953-7143f648c9ae-config-data\") pod \"rabbitmq-server-0\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " pod="openstack/rabbitmq-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.795527 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e762572f-d5b9-462e-9953-7143f648c9ae-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " pod="openstack/rabbitmq-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.796016 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e762572f-d5b9-462e-9953-7143f648c9ae-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " pod="openstack/rabbitmq-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.797567 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e762572f-d5b9-462e-9953-7143f648c9ae-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " pod="openstack/rabbitmq-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.808769 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6qkc\" (UniqueName: \"kubernetes.io/projected/e762572f-d5b9-462e-9953-7143f648c9ae-kube-api-access-c6qkc\") pod \"rabbitmq-server-0\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " pod="openstack/rabbitmq-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.813627 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e762572f-d5b9-462e-9953-7143f648c9ae-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " pod="openstack/rabbitmq-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.839857 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " pod="openstack/rabbitmq-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.872635 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77795d58f5-kh2gz"] Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.888628 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.894018 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.898944 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.903707 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.903942 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.904420 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-rk8pz" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.905264 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.905990 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.906114 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.907848 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.918021 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.994498 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.994560 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ce587847-47c2-41de-95d1-c9f8ab88961e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.994608 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ce587847-47c2-41de-95d1-c9f8ab88961e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.994640 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ce587847-47c2-41de-95d1-c9f8ab88961e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.994669 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ce587847-47c2-41de-95d1-c9f8ab88961e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.994868 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ce587847-47c2-41de-95d1-c9f8ab88961e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.994967 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ce587847-47c2-41de-95d1-c9f8ab88961e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.995023 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqw5k\" (UniqueName: \"kubernetes.io/projected/ce587847-47c2-41de-95d1-c9f8ab88961e-kube-api-access-tqw5k\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.995142 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ce587847-47c2-41de-95d1-c9f8ab88961e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.995211 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ce587847-47c2-41de-95d1-c9f8ab88961e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:50:37 crc kubenswrapper[4787]: I1001 09:50:37.995419 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ce587847-47c2-41de-95d1-c9f8ab88961e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:50:38 crc kubenswrapper[4787]: I1001 09:50:38.096749 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ce587847-47c2-41de-95d1-c9f8ab88961e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:50:38 crc kubenswrapper[4787]: I1001 09:50:38.097168 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:50:38 crc kubenswrapper[4787]: I1001 09:50:38.097194 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ce587847-47c2-41de-95d1-c9f8ab88961e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:50:38 crc kubenswrapper[4787]: I1001 09:50:38.097355 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:50:38 crc kubenswrapper[4787]: I1001 09:50:38.102376 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ce587847-47c2-41de-95d1-c9f8ab88961e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:50:38 crc kubenswrapper[4787]: I1001 09:50:38.102547 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ce587847-47c2-41de-95d1-c9f8ab88961e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:50:38 crc kubenswrapper[4787]: I1001 09:50:38.102593 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ce587847-47c2-41de-95d1-c9f8ab88961e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:50:38 crc kubenswrapper[4787]: I1001 09:50:38.102653 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ce587847-47c2-41de-95d1-c9f8ab88961e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:50:38 crc kubenswrapper[4787]: I1001 09:50:38.102694 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ce587847-47c2-41de-95d1-c9f8ab88961e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:50:38 crc kubenswrapper[4787]: I1001 09:50:38.102722 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqw5k\" (UniqueName: \"kubernetes.io/projected/ce587847-47c2-41de-95d1-c9f8ab88961e-kube-api-access-tqw5k\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:50:38 crc kubenswrapper[4787]: I1001 09:50:38.103541 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ce587847-47c2-41de-95d1-c9f8ab88961e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:50:38 crc kubenswrapper[4787]: I1001 09:50:38.103812 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ce587847-47c2-41de-95d1-c9f8ab88961e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:50:38 crc kubenswrapper[4787]: I1001 09:50:38.104312 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ce587847-47c2-41de-95d1-c9f8ab88961e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:50:38 crc kubenswrapper[4787]: I1001 09:50:38.104411 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ce587847-47c2-41de-95d1-c9f8ab88961e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:50:38 crc kubenswrapper[4787]: I1001 09:50:38.104926 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ce587847-47c2-41de-95d1-c9f8ab88961e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:50:38 crc kubenswrapper[4787]: I1001 09:50:38.109149 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ce587847-47c2-41de-95d1-c9f8ab88961e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:50:38 crc kubenswrapper[4787]: I1001 09:50:38.111025 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ce587847-47c2-41de-95d1-c9f8ab88961e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:50:38 crc kubenswrapper[4787]: I1001 09:50:38.111123 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ce587847-47c2-41de-95d1-c9f8ab88961e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:50:38 crc kubenswrapper[4787]: I1001 09:50:38.111532 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ce587847-47c2-41de-95d1-c9f8ab88961e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:50:38 crc kubenswrapper[4787]: I1001 09:50:38.112459 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ce587847-47c2-41de-95d1-c9f8ab88961e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:50:38 crc kubenswrapper[4787]: I1001 09:50:38.119412 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ce587847-47c2-41de-95d1-c9f8ab88961e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:50:38 crc kubenswrapper[4787]: I1001 09:50:38.132205 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqw5k\" (UniqueName: \"kubernetes.io/projected/ce587847-47c2-41de-95d1-c9f8ab88961e-kube-api-access-tqw5k\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:50:38 crc kubenswrapper[4787]: I1001 09:50:38.132529 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:50:38 crc kubenswrapper[4787]: I1001 09:50:38.233661 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b6f94bdfc-jdsn9" event={"ID":"a5c38d07-c1a7-46fd-bcce-1680eff35b1f","Type":"ContainerStarted","Data":"b601caa8d6560d005c9b5fd5d227856c95031c41c0f73e2e9ef3ac890b073f64"} Oct 01 09:50:38 crc kubenswrapper[4787]: I1001 09:50:38.235673 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77795d58f5-kh2gz" event={"ID":"105b8088-a178-492b-aefd-20a5e6ad0dc8","Type":"ContainerStarted","Data":"f8ab9a21de90c12090d0a833ff82757b32680470016b00c2d78164bca607be41"} Oct 01 09:50:38 crc kubenswrapper[4787]: I1001 09:50:38.250219 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:50:38 crc kubenswrapper[4787]: I1001 09:50:38.416801 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 09:50:38 crc kubenswrapper[4787]: I1001 09:50:38.834862 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 09:50:39 crc kubenswrapper[4787]: I1001 09:50:39.248834 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e762572f-d5b9-462e-9953-7143f648c9ae","Type":"ContainerStarted","Data":"1fa050ebea47947b00534e1bcd8a7e26894e48cad6c5ba6f3fba7bfb5bc0d198"} Oct 01 09:50:39 crc kubenswrapper[4787]: I1001 09:50:39.251237 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ce587847-47c2-41de-95d1-c9f8ab88961e","Type":"ContainerStarted","Data":"12dc2239ff3366a6d3e4d8d866586456447e31020fbee21027bf57f38d6be118"} Oct 01 09:50:39 crc kubenswrapper[4787]: I1001 09:50:39.450137 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 01 09:50:39 crc kubenswrapper[4787]: I1001 09:50:39.451496 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 01 09:50:39 crc kubenswrapper[4787]: I1001 09:50:39.457885 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 01 09:50:39 crc kubenswrapper[4787]: I1001 09:50:39.458114 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 01 09:50:39 crc kubenswrapper[4787]: I1001 09:50:39.459460 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 01 09:50:39 crc kubenswrapper[4787]: I1001 09:50:39.459551 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-cckf8" Oct 01 09:50:39 crc kubenswrapper[4787]: I1001 09:50:39.460878 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 01 09:50:39 crc kubenswrapper[4787]: I1001 09:50:39.466206 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 01 09:50:39 crc kubenswrapper[4787]: I1001 09:50:39.468137 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 01 09:50:39 crc kubenswrapper[4787]: I1001 09:50:39.545342 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c-config-data-default\") pod \"openstack-galera-0\" (UID: \"37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c\") " pod="openstack/openstack-galera-0" Oct 01 09:50:39 crc kubenswrapper[4787]: I1001 09:50:39.545451 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c-secrets\") pod \"openstack-galera-0\" (UID: \"37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c\") " pod="openstack/openstack-galera-0" Oct 01 09:50:39 crc kubenswrapper[4787]: I1001 09:50:39.545572 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c\") " pod="openstack/openstack-galera-0" Oct 01 09:50:39 crc kubenswrapper[4787]: I1001 09:50:39.545701 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c\") " pod="openstack/openstack-galera-0" Oct 01 09:50:39 crc kubenswrapper[4787]: I1001 09:50:39.545747 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kbn6\" (UniqueName: \"kubernetes.io/projected/37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c-kube-api-access-8kbn6\") pod \"openstack-galera-0\" (UID: \"37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c\") " pod="openstack/openstack-galera-0" Oct 01 09:50:39 crc kubenswrapper[4787]: I1001 09:50:39.545831 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c-operator-scripts\") pod \"openstack-galera-0\" (UID: \"37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c\") " pod="openstack/openstack-galera-0" Oct 01 09:50:39 crc kubenswrapper[4787]: I1001 09:50:39.545884 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c\") " pod="openstack/openstack-galera-0" Oct 01 09:50:39 crc kubenswrapper[4787]: I1001 09:50:39.545901 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c-kolla-config\") pod \"openstack-galera-0\" (UID: \"37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c\") " pod="openstack/openstack-galera-0" Oct 01 09:50:39 crc kubenswrapper[4787]: I1001 09:50:39.546008 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c-config-data-generated\") pod \"openstack-galera-0\" (UID: \"37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c\") " pod="openstack/openstack-galera-0" Oct 01 09:50:39 crc kubenswrapper[4787]: I1001 09:50:39.647401 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c\") " pod="openstack/openstack-galera-0" Oct 01 09:50:39 crc kubenswrapper[4787]: I1001 09:50:39.647488 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kbn6\" (UniqueName: \"kubernetes.io/projected/37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c-kube-api-access-8kbn6\") pod \"openstack-galera-0\" (UID: \"37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c\") " pod="openstack/openstack-galera-0" Oct 01 09:50:39 crc kubenswrapper[4787]: I1001 09:50:39.647537 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c-operator-scripts\") pod \"openstack-galera-0\" (UID: \"37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c\") " pod="openstack/openstack-galera-0" Oct 01 09:50:39 crc kubenswrapper[4787]: I1001 09:50:39.647606 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c\") " pod="openstack/openstack-galera-0" Oct 01 09:50:39 crc kubenswrapper[4787]: I1001 09:50:39.647627 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c-kolla-config\") pod \"openstack-galera-0\" (UID: \"37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c\") " pod="openstack/openstack-galera-0" Oct 01 09:50:39 crc kubenswrapper[4787]: I1001 09:50:39.647655 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c-config-data-generated\") pod \"openstack-galera-0\" (UID: \"37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c\") " pod="openstack/openstack-galera-0" Oct 01 09:50:39 crc kubenswrapper[4787]: I1001 09:50:39.647742 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c-config-data-default\") pod \"openstack-galera-0\" (UID: \"37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c\") " pod="openstack/openstack-galera-0" Oct 01 09:50:39 crc kubenswrapper[4787]: I1001 09:50:39.647763 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c-secrets\") pod \"openstack-galera-0\" (UID: \"37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c\") " pod="openstack/openstack-galera-0" Oct 01 09:50:39 crc kubenswrapper[4787]: I1001 09:50:39.647831 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c\") " pod="openstack/openstack-galera-0" Oct 01 09:50:39 crc kubenswrapper[4787]: I1001 09:50:39.647931 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/openstack-galera-0" Oct 01 09:50:39 crc kubenswrapper[4787]: I1001 09:50:39.648909 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c-kolla-config\") pod \"openstack-galera-0\" (UID: \"37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c\") " pod="openstack/openstack-galera-0" Oct 01 09:50:39 crc kubenswrapper[4787]: I1001 09:50:39.649945 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c-operator-scripts\") pod \"openstack-galera-0\" (UID: \"37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c\") " pod="openstack/openstack-galera-0" Oct 01 09:50:39 crc kubenswrapper[4787]: I1001 09:50:39.650351 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c-config-data-generated\") pod \"openstack-galera-0\" (UID: \"37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c\") " pod="openstack/openstack-galera-0" Oct 01 09:50:39 crc kubenswrapper[4787]: I1001 09:50:39.650861 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c-config-data-default\") pod \"openstack-galera-0\" (UID: \"37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c\") " pod="openstack/openstack-galera-0" Oct 01 09:50:39 crc kubenswrapper[4787]: I1001 09:50:39.656056 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c-secrets\") pod \"openstack-galera-0\" (UID: \"37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c\") " pod="openstack/openstack-galera-0" Oct 01 09:50:39 crc kubenswrapper[4787]: I1001 09:50:39.661662 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c\") " pod="openstack/openstack-galera-0" Oct 01 09:50:39 crc kubenswrapper[4787]: I1001 09:50:39.673094 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kbn6\" (UniqueName: \"kubernetes.io/projected/37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c-kube-api-access-8kbn6\") pod \"openstack-galera-0\" (UID: \"37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c\") " pod="openstack/openstack-galera-0" Oct 01 09:50:39 crc kubenswrapper[4787]: I1001 09:50:39.686903 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c\") " pod="openstack/openstack-galera-0" Oct 01 09:50:39 crc kubenswrapper[4787]: I1001 09:50:39.712319 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c\") " pod="openstack/openstack-galera-0" Oct 01 09:50:39 crc kubenswrapper[4787]: I1001 09:50:39.799518 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.541702 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.543027 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.545725 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-5whrd" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.545853 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.546000 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.546258 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.579859 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.670330 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7924a2ac-f411-42d6-8b23-f22faf58351a-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"7924a2ac-f411-42d6-8b23-f22faf58351a\") " pod="openstack/openstack-cell1-galera-0" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.670827 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7924a2ac-f411-42d6-8b23-f22faf58351a-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"7924a2ac-f411-42d6-8b23-f22faf58351a\") " pod="openstack/openstack-cell1-galera-0" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.670996 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9gxd\" (UniqueName: \"kubernetes.io/projected/7924a2ac-f411-42d6-8b23-f22faf58351a-kube-api-access-f9gxd\") pod \"openstack-cell1-galera-0\" (UID: \"7924a2ac-f411-42d6-8b23-f22faf58351a\") " pod="openstack/openstack-cell1-galera-0" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.671128 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/7924a2ac-f411-42d6-8b23-f22faf58351a-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"7924a2ac-f411-42d6-8b23-f22faf58351a\") " pod="openstack/openstack-cell1-galera-0" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.671205 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7924a2ac-f411-42d6-8b23-f22faf58351a-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"7924a2ac-f411-42d6-8b23-f22faf58351a\") " pod="openstack/openstack-cell1-galera-0" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.671256 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7924a2ac-f411-42d6-8b23-f22faf58351a-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"7924a2ac-f411-42d6-8b23-f22faf58351a\") " pod="openstack/openstack-cell1-galera-0" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.671284 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7924a2ac-f411-42d6-8b23-f22faf58351a-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"7924a2ac-f411-42d6-8b23-f22faf58351a\") " pod="openstack/openstack-cell1-galera-0" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.671345 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7924a2ac-f411-42d6-8b23-f22faf58351a\") " pod="openstack/openstack-cell1-galera-0" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.671374 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7924a2ac-f411-42d6-8b23-f22faf58351a-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"7924a2ac-f411-42d6-8b23-f22faf58351a\") " pod="openstack/openstack-cell1-galera-0" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.773124 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7924a2ac-f411-42d6-8b23-f22faf58351a-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"7924a2ac-f411-42d6-8b23-f22faf58351a\") " pod="openstack/openstack-cell1-galera-0" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.773184 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9gxd\" (UniqueName: \"kubernetes.io/projected/7924a2ac-f411-42d6-8b23-f22faf58351a-kube-api-access-f9gxd\") pod \"openstack-cell1-galera-0\" (UID: \"7924a2ac-f411-42d6-8b23-f22faf58351a\") " pod="openstack/openstack-cell1-galera-0" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.773214 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/7924a2ac-f411-42d6-8b23-f22faf58351a-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"7924a2ac-f411-42d6-8b23-f22faf58351a\") " pod="openstack/openstack-cell1-galera-0" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.773283 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7924a2ac-f411-42d6-8b23-f22faf58351a-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"7924a2ac-f411-42d6-8b23-f22faf58351a\") " pod="openstack/openstack-cell1-galera-0" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.773303 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7924a2ac-f411-42d6-8b23-f22faf58351a-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"7924a2ac-f411-42d6-8b23-f22faf58351a\") " pod="openstack/openstack-cell1-galera-0" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.773320 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7924a2ac-f411-42d6-8b23-f22faf58351a-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"7924a2ac-f411-42d6-8b23-f22faf58351a\") " pod="openstack/openstack-cell1-galera-0" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.773340 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7924a2ac-f411-42d6-8b23-f22faf58351a\") " pod="openstack/openstack-cell1-galera-0" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.773357 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7924a2ac-f411-42d6-8b23-f22faf58351a-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"7924a2ac-f411-42d6-8b23-f22faf58351a\") " pod="openstack/openstack-cell1-galera-0" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.773441 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7924a2ac-f411-42d6-8b23-f22faf58351a-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"7924a2ac-f411-42d6-8b23-f22faf58351a\") " pod="openstack/openstack-cell1-galera-0" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.773977 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7924a2ac-f411-42d6-8b23-f22faf58351a-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"7924a2ac-f411-42d6-8b23-f22faf58351a\") " pod="openstack/openstack-cell1-galera-0" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.774251 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7924a2ac-f411-42d6-8b23-f22faf58351a\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/openstack-cell1-galera-0" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.774633 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7924a2ac-f411-42d6-8b23-f22faf58351a-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"7924a2ac-f411-42d6-8b23-f22faf58351a\") " pod="openstack/openstack-cell1-galera-0" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.776343 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7924a2ac-f411-42d6-8b23-f22faf58351a-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"7924a2ac-f411-42d6-8b23-f22faf58351a\") " pod="openstack/openstack-cell1-galera-0" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.778735 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7924a2ac-f411-42d6-8b23-f22faf58351a-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"7924a2ac-f411-42d6-8b23-f22faf58351a\") " pod="openstack/openstack-cell1-galera-0" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.779258 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/7924a2ac-f411-42d6-8b23-f22faf58351a-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"7924a2ac-f411-42d6-8b23-f22faf58351a\") " pod="openstack/openstack-cell1-galera-0" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.779401 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7924a2ac-f411-42d6-8b23-f22faf58351a-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"7924a2ac-f411-42d6-8b23-f22faf58351a\") " pod="openstack/openstack-cell1-galera-0" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.780218 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7924a2ac-f411-42d6-8b23-f22faf58351a-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"7924a2ac-f411-42d6-8b23-f22faf58351a\") " pod="openstack/openstack-cell1-galera-0" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.807938 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9gxd\" (UniqueName: \"kubernetes.io/projected/7924a2ac-f411-42d6-8b23-f22faf58351a-kube-api-access-f9gxd\") pod \"openstack-cell1-galera-0\" (UID: \"7924a2ac-f411-42d6-8b23-f22faf58351a\") " pod="openstack/openstack-cell1-galera-0" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.845846 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7924a2ac-f411-42d6-8b23-f22faf58351a\") " pod="openstack/openstack-cell1-galera-0" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.870004 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.883768 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.886660 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.886909 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-h62kg" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.887352 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.894432 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.900404 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.979154 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/7137d092-ff92-4b8d-84c5-9d9d269ddcc7-memcached-tls-certs\") pod \"memcached-0\" (UID: \"7137d092-ff92-4b8d-84c5-9d9d269ddcc7\") " pod="openstack/memcached-0" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.979236 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7137d092-ff92-4b8d-84c5-9d9d269ddcc7-kolla-config\") pod \"memcached-0\" (UID: \"7137d092-ff92-4b8d-84c5-9d9d269ddcc7\") " pod="openstack/memcached-0" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.979387 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7137d092-ff92-4b8d-84c5-9d9d269ddcc7-combined-ca-bundle\") pod \"memcached-0\" (UID: \"7137d092-ff92-4b8d-84c5-9d9d269ddcc7\") " pod="openstack/memcached-0" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.979781 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r84mp\" (UniqueName: \"kubernetes.io/projected/7137d092-ff92-4b8d-84c5-9d9d269ddcc7-kube-api-access-r84mp\") pod \"memcached-0\" (UID: \"7137d092-ff92-4b8d-84c5-9d9d269ddcc7\") " pod="openstack/memcached-0" Oct 01 09:50:40 crc kubenswrapper[4787]: I1001 09:50:40.979852 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7137d092-ff92-4b8d-84c5-9d9d269ddcc7-config-data\") pod \"memcached-0\" (UID: \"7137d092-ff92-4b8d-84c5-9d9d269ddcc7\") " pod="openstack/memcached-0" Oct 01 09:50:41 crc kubenswrapper[4787]: I1001 09:50:41.081286 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r84mp\" (UniqueName: \"kubernetes.io/projected/7137d092-ff92-4b8d-84c5-9d9d269ddcc7-kube-api-access-r84mp\") pod \"memcached-0\" (UID: \"7137d092-ff92-4b8d-84c5-9d9d269ddcc7\") " pod="openstack/memcached-0" Oct 01 09:50:41 crc kubenswrapper[4787]: I1001 09:50:41.081373 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7137d092-ff92-4b8d-84c5-9d9d269ddcc7-config-data\") pod \"memcached-0\" (UID: \"7137d092-ff92-4b8d-84c5-9d9d269ddcc7\") " pod="openstack/memcached-0" Oct 01 09:50:41 crc kubenswrapper[4787]: I1001 09:50:41.081402 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/7137d092-ff92-4b8d-84c5-9d9d269ddcc7-memcached-tls-certs\") pod \"memcached-0\" (UID: \"7137d092-ff92-4b8d-84c5-9d9d269ddcc7\") " pod="openstack/memcached-0" Oct 01 09:50:41 crc kubenswrapper[4787]: I1001 09:50:41.081471 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7137d092-ff92-4b8d-84c5-9d9d269ddcc7-kolla-config\") pod \"memcached-0\" (UID: \"7137d092-ff92-4b8d-84c5-9d9d269ddcc7\") " pod="openstack/memcached-0" Oct 01 09:50:41 crc kubenswrapper[4787]: I1001 09:50:41.081506 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7137d092-ff92-4b8d-84c5-9d9d269ddcc7-combined-ca-bundle\") pod \"memcached-0\" (UID: \"7137d092-ff92-4b8d-84c5-9d9d269ddcc7\") " pod="openstack/memcached-0" Oct 01 09:50:41 crc kubenswrapper[4787]: I1001 09:50:41.082394 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7137d092-ff92-4b8d-84c5-9d9d269ddcc7-kolla-config\") pod \"memcached-0\" (UID: \"7137d092-ff92-4b8d-84c5-9d9d269ddcc7\") " pod="openstack/memcached-0" Oct 01 09:50:41 crc kubenswrapper[4787]: I1001 09:50:41.082887 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7137d092-ff92-4b8d-84c5-9d9d269ddcc7-config-data\") pod \"memcached-0\" (UID: \"7137d092-ff92-4b8d-84c5-9d9d269ddcc7\") " pod="openstack/memcached-0" Oct 01 09:50:41 crc kubenswrapper[4787]: I1001 09:50:41.085261 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/7137d092-ff92-4b8d-84c5-9d9d269ddcc7-memcached-tls-certs\") pod \"memcached-0\" (UID: \"7137d092-ff92-4b8d-84c5-9d9d269ddcc7\") " pod="openstack/memcached-0" Oct 01 09:50:41 crc kubenswrapper[4787]: I1001 09:50:41.089991 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7137d092-ff92-4b8d-84c5-9d9d269ddcc7-combined-ca-bundle\") pod \"memcached-0\" (UID: \"7137d092-ff92-4b8d-84c5-9d9d269ddcc7\") " pod="openstack/memcached-0" Oct 01 09:50:41 crc kubenswrapper[4787]: I1001 09:50:41.108796 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r84mp\" (UniqueName: \"kubernetes.io/projected/7137d092-ff92-4b8d-84c5-9d9d269ddcc7-kube-api-access-r84mp\") pod \"memcached-0\" (UID: \"7137d092-ff92-4b8d-84c5-9d9d269ddcc7\") " pod="openstack/memcached-0" Oct 01 09:50:41 crc kubenswrapper[4787]: I1001 09:50:41.232955 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 01 09:50:41 crc kubenswrapper[4787]: I1001 09:50:41.251050 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:50:41 crc kubenswrapper[4787]: I1001 09:50:41.251141 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:50:42 crc kubenswrapper[4787]: I1001 09:50:42.887498 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 09:50:42 crc kubenswrapper[4787]: I1001 09:50:42.888915 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 01 09:50:42 crc kubenswrapper[4787]: I1001 09:50:42.892481 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-j9stt" Oct 01 09:50:42 crc kubenswrapper[4787]: I1001 09:50:42.900666 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 09:50:42 crc kubenswrapper[4787]: I1001 09:50:42.915259 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wj8f6\" (UniqueName: \"kubernetes.io/projected/469f4dc6-d176-4497-ac69-6130a375737b-kube-api-access-wj8f6\") pod \"kube-state-metrics-0\" (UID: \"469f4dc6-d176-4497-ac69-6130a375737b\") " pod="openstack/kube-state-metrics-0" Oct 01 09:50:43 crc kubenswrapper[4787]: I1001 09:50:43.016295 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wj8f6\" (UniqueName: \"kubernetes.io/projected/469f4dc6-d176-4497-ac69-6130a375737b-kube-api-access-wj8f6\") pod \"kube-state-metrics-0\" (UID: \"469f4dc6-d176-4497-ac69-6130a375737b\") " pod="openstack/kube-state-metrics-0" Oct 01 09:50:43 crc kubenswrapper[4787]: I1001 09:50:43.037142 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wj8f6\" (UniqueName: \"kubernetes.io/projected/469f4dc6-d176-4497-ac69-6130a375737b-kube-api-access-wj8f6\") pod \"kube-state-metrics-0\" (UID: \"469f4dc6-d176-4497-ac69-6130a375737b\") " pod="openstack/kube-state-metrics-0" Oct 01 09:50:43 crc kubenswrapper[4787]: I1001 09:50:43.216425 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 01 09:50:44 crc kubenswrapper[4787]: I1001 09:50:44.647581 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 01 09:50:46 crc kubenswrapper[4787]: I1001 09:50:46.851950 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 01 09:50:46 crc kubenswrapper[4787]: I1001 09:50:46.854537 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 01 09:50:46 crc kubenswrapper[4787]: I1001 09:50:46.860277 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 01 09:50:46 crc kubenswrapper[4787]: I1001 09:50:46.860733 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 01 09:50:46 crc kubenswrapper[4787]: I1001 09:50:46.861007 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-srml8" Oct 01 09:50:46 crc kubenswrapper[4787]: I1001 09:50:46.861023 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 01 09:50:46 crc kubenswrapper[4787]: I1001 09:50:46.861023 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 01 09:50:46 crc kubenswrapper[4787]: I1001 09:50:46.867706 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 01 09:50:46 crc kubenswrapper[4787]: I1001 09:50:46.992879 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee50aeec-8875-45c2-80e2-7121480ae89f-config\") pod \"ovsdbserver-nb-0\" (UID: \"ee50aeec-8875-45c2-80e2-7121480ae89f\") " pod="openstack/ovsdbserver-nb-0" Oct 01 09:50:46 crc kubenswrapper[4787]: I1001 09:50:46.993010 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee50aeec-8875-45c2-80e2-7121480ae89f-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"ee50aeec-8875-45c2-80e2-7121480ae89f\") " pod="openstack/ovsdbserver-nb-0" Oct 01 09:50:46 crc kubenswrapper[4787]: I1001 09:50:46.993056 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee50aeec-8875-45c2-80e2-7121480ae89f-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ee50aeec-8875-45c2-80e2-7121480ae89f\") " pod="openstack/ovsdbserver-nb-0" Oct 01 09:50:46 crc kubenswrapper[4787]: I1001 09:50:46.993235 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6g792\" (UniqueName: \"kubernetes.io/projected/ee50aeec-8875-45c2-80e2-7121480ae89f-kube-api-access-6g792\") pod \"ovsdbserver-nb-0\" (UID: \"ee50aeec-8875-45c2-80e2-7121480ae89f\") " pod="openstack/ovsdbserver-nb-0" Oct 01 09:50:46 crc kubenswrapper[4787]: I1001 09:50:46.993259 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"ee50aeec-8875-45c2-80e2-7121480ae89f\") " pod="openstack/ovsdbserver-nb-0" Oct 01 09:50:46 crc kubenswrapper[4787]: I1001 09:50:46.993302 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee50aeec-8875-45c2-80e2-7121480ae89f-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"ee50aeec-8875-45c2-80e2-7121480ae89f\") " pod="openstack/ovsdbserver-nb-0" Oct 01 09:50:46 crc kubenswrapper[4787]: I1001 09:50:46.993332 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee50aeec-8875-45c2-80e2-7121480ae89f-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ee50aeec-8875-45c2-80e2-7121480ae89f\") " pod="openstack/ovsdbserver-nb-0" Oct 01 09:50:46 crc kubenswrapper[4787]: I1001 09:50:46.993383 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ee50aeec-8875-45c2-80e2-7121480ae89f-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"ee50aeec-8875-45c2-80e2-7121480ae89f\") " pod="openstack/ovsdbserver-nb-0" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.094466 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee50aeec-8875-45c2-80e2-7121480ae89f-config\") pod \"ovsdbserver-nb-0\" (UID: \"ee50aeec-8875-45c2-80e2-7121480ae89f\") " pod="openstack/ovsdbserver-nb-0" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.094538 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee50aeec-8875-45c2-80e2-7121480ae89f-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"ee50aeec-8875-45c2-80e2-7121480ae89f\") " pod="openstack/ovsdbserver-nb-0" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.094577 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee50aeec-8875-45c2-80e2-7121480ae89f-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ee50aeec-8875-45c2-80e2-7121480ae89f\") " pod="openstack/ovsdbserver-nb-0" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.094607 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6g792\" (UniqueName: \"kubernetes.io/projected/ee50aeec-8875-45c2-80e2-7121480ae89f-kube-api-access-6g792\") pod \"ovsdbserver-nb-0\" (UID: \"ee50aeec-8875-45c2-80e2-7121480ae89f\") " pod="openstack/ovsdbserver-nb-0" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.094629 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"ee50aeec-8875-45c2-80e2-7121480ae89f\") " pod="openstack/ovsdbserver-nb-0" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.094658 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee50aeec-8875-45c2-80e2-7121480ae89f-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"ee50aeec-8875-45c2-80e2-7121480ae89f\") " pod="openstack/ovsdbserver-nb-0" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.094688 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee50aeec-8875-45c2-80e2-7121480ae89f-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ee50aeec-8875-45c2-80e2-7121480ae89f\") " pod="openstack/ovsdbserver-nb-0" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.094727 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ee50aeec-8875-45c2-80e2-7121480ae89f-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"ee50aeec-8875-45c2-80e2-7121480ae89f\") " pod="openstack/ovsdbserver-nb-0" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.095068 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"ee50aeec-8875-45c2-80e2-7121480ae89f\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/ovsdbserver-nb-0" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.095342 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee50aeec-8875-45c2-80e2-7121480ae89f-config\") pod \"ovsdbserver-nb-0\" (UID: \"ee50aeec-8875-45c2-80e2-7121480ae89f\") " pod="openstack/ovsdbserver-nb-0" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.095097 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ee50aeec-8875-45c2-80e2-7121480ae89f-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"ee50aeec-8875-45c2-80e2-7121480ae89f\") " pod="openstack/ovsdbserver-nb-0" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.096003 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee50aeec-8875-45c2-80e2-7121480ae89f-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"ee50aeec-8875-45c2-80e2-7121480ae89f\") " pod="openstack/ovsdbserver-nb-0" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.101723 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee50aeec-8875-45c2-80e2-7121480ae89f-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"ee50aeec-8875-45c2-80e2-7121480ae89f\") " pod="openstack/ovsdbserver-nb-0" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.105097 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee50aeec-8875-45c2-80e2-7121480ae89f-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ee50aeec-8875-45c2-80e2-7121480ae89f\") " pod="openstack/ovsdbserver-nb-0" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.115241 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6g792\" (UniqueName: \"kubernetes.io/projected/ee50aeec-8875-45c2-80e2-7121480ae89f-kube-api-access-6g792\") pod \"ovsdbserver-nb-0\" (UID: \"ee50aeec-8875-45c2-80e2-7121480ae89f\") " pod="openstack/ovsdbserver-nb-0" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.117570 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee50aeec-8875-45c2-80e2-7121480ae89f-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ee50aeec-8875-45c2-80e2-7121480ae89f\") " pod="openstack/ovsdbserver-nb-0" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.131892 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"ee50aeec-8875-45c2-80e2-7121480ae89f\") " pod="openstack/ovsdbserver-nb-0" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.187093 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.447039 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-l8fqd"] Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.448634 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-l8fqd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.450600 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.450992 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.451331 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-qs9jj" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.459893 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-l8fqd"] Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.468520 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-6htcd"] Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.470726 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-6htcd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.477472 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-6htcd"] Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.500412 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/36eb99a8-609a-4af2-9fa7-d0051806659e-ovn-controller-tls-certs\") pod \"ovn-controller-l8fqd\" (UID: \"36eb99a8-609a-4af2-9fa7-d0051806659e\") " pod="openstack/ovn-controller-l8fqd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.500487 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/36eb99a8-609a-4af2-9fa7-d0051806659e-var-log-ovn\") pod \"ovn-controller-l8fqd\" (UID: \"36eb99a8-609a-4af2-9fa7-d0051806659e\") " pod="openstack/ovn-controller-l8fqd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.500521 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/36eb99a8-609a-4af2-9fa7-d0051806659e-var-run\") pod \"ovn-controller-l8fqd\" (UID: \"36eb99a8-609a-4af2-9fa7-d0051806659e\") " pod="openstack/ovn-controller-l8fqd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.500793 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/36eb99a8-609a-4af2-9fa7-d0051806659e-var-run-ovn\") pod \"ovn-controller-l8fqd\" (UID: \"36eb99a8-609a-4af2-9fa7-d0051806659e\") " pod="openstack/ovn-controller-l8fqd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.500897 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gv9lx\" (UniqueName: \"kubernetes.io/projected/36eb99a8-609a-4af2-9fa7-d0051806659e-kube-api-access-gv9lx\") pod \"ovn-controller-l8fqd\" (UID: \"36eb99a8-609a-4af2-9fa7-d0051806659e\") " pod="openstack/ovn-controller-l8fqd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.500988 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/36eb99a8-609a-4af2-9fa7-d0051806659e-scripts\") pod \"ovn-controller-l8fqd\" (UID: \"36eb99a8-609a-4af2-9fa7-d0051806659e\") " pod="openstack/ovn-controller-l8fqd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.501016 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36eb99a8-609a-4af2-9fa7-d0051806659e-combined-ca-bundle\") pod \"ovn-controller-l8fqd\" (UID: \"36eb99a8-609a-4af2-9fa7-d0051806659e\") " pod="openstack/ovn-controller-l8fqd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.602890 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/36eb99a8-609a-4af2-9fa7-d0051806659e-scripts\") pod \"ovn-controller-l8fqd\" (UID: \"36eb99a8-609a-4af2-9fa7-d0051806659e\") " pod="openstack/ovn-controller-l8fqd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.602947 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36eb99a8-609a-4af2-9fa7-d0051806659e-combined-ca-bundle\") pod \"ovn-controller-l8fqd\" (UID: \"36eb99a8-609a-4af2-9fa7-d0051806659e\") " pod="openstack/ovn-controller-l8fqd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.602976 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/948a10c7-fb61-4110-a08e-b2fd42cb842a-var-log\") pod \"ovn-controller-ovs-6htcd\" (UID: \"948a10c7-fb61-4110-a08e-b2fd42cb842a\") " pod="openstack/ovn-controller-ovs-6htcd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.603008 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5knh6\" (UniqueName: \"kubernetes.io/projected/948a10c7-fb61-4110-a08e-b2fd42cb842a-kube-api-access-5knh6\") pod \"ovn-controller-ovs-6htcd\" (UID: \"948a10c7-fb61-4110-a08e-b2fd42cb842a\") " pod="openstack/ovn-controller-ovs-6htcd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.603051 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/948a10c7-fb61-4110-a08e-b2fd42cb842a-scripts\") pod \"ovn-controller-ovs-6htcd\" (UID: \"948a10c7-fb61-4110-a08e-b2fd42cb842a\") " pod="openstack/ovn-controller-ovs-6htcd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.603087 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/36eb99a8-609a-4af2-9fa7-d0051806659e-ovn-controller-tls-certs\") pod \"ovn-controller-l8fqd\" (UID: \"36eb99a8-609a-4af2-9fa7-d0051806659e\") " pod="openstack/ovn-controller-l8fqd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.603179 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/36eb99a8-609a-4af2-9fa7-d0051806659e-var-log-ovn\") pod \"ovn-controller-l8fqd\" (UID: \"36eb99a8-609a-4af2-9fa7-d0051806659e\") " pod="openstack/ovn-controller-l8fqd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.603202 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/948a10c7-fb61-4110-a08e-b2fd42cb842a-var-run\") pod \"ovn-controller-ovs-6htcd\" (UID: \"948a10c7-fb61-4110-a08e-b2fd42cb842a\") " pod="openstack/ovn-controller-ovs-6htcd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.603223 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/948a10c7-fb61-4110-a08e-b2fd42cb842a-etc-ovs\") pod \"ovn-controller-ovs-6htcd\" (UID: \"948a10c7-fb61-4110-a08e-b2fd42cb842a\") " pod="openstack/ovn-controller-ovs-6htcd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.603245 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/36eb99a8-609a-4af2-9fa7-d0051806659e-var-run\") pod \"ovn-controller-l8fqd\" (UID: \"36eb99a8-609a-4af2-9fa7-d0051806659e\") " pod="openstack/ovn-controller-l8fqd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.603270 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/36eb99a8-609a-4af2-9fa7-d0051806659e-var-run-ovn\") pod \"ovn-controller-l8fqd\" (UID: \"36eb99a8-609a-4af2-9fa7-d0051806659e\") " pod="openstack/ovn-controller-l8fqd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.603356 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/948a10c7-fb61-4110-a08e-b2fd42cb842a-var-lib\") pod \"ovn-controller-ovs-6htcd\" (UID: \"948a10c7-fb61-4110-a08e-b2fd42cb842a\") " pod="openstack/ovn-controller-ovs-6htcd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.603527 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gv9lx\" (UniqueName: \"kubernetes.io/projected/36eb99a8-609a-4af2-9fa7-d0051806659e-kube-api-access-gv9lx\") pod \"ovn-controller-l8fqd\" (UID: \"36eb99a8-609a-4af2-9fa7-d0051806659e\") " pod="openstack/ovn-controller-l8fqd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.603778 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/36eb99a8-609a-4af2-9fa7-d0051806659e-var-log-ovn\") pod \"ovn-controller-l8fqd\" (UID: \"36eb99a8-609a-4af2-9fa7-d0051806659e\") " pod="openstack/ovn-controller-l8fqd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.603857 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/36eb99a8-609a-4af2-9fa7-d0051806659e-var-run-ovn\") pod \"ovn-controller-l8fqd\" (UID: \"36eb99a8-609a-4af2-9fa7-d0051806659e\") " pod="openstack/ovn-controller-l8fqd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.603925 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/36eb99a8-609a-4af2-9fa7-d0051806659e-var-run\") pod \"ovn-controller-l8fqd\" (UID: \"36eb99a8-609a-4af2-9fa7-d0051806659e\") " pod="openstack/ovn-controller-l8fqd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.606459 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/36eb99a8-609a-4af2-9fa7-d0051806659e-ovn-controller-tls-certs\") pod \"ovn-controller-l8fqd\" (UID: \"36eb99a8-609a-4af2-9fa7-d0051806659e\") " pod="openstack/ovn-controller-l8fqd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.607190 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36eb99a8-609a-4af2-9fa7-d0051806659e-combined-ca-bundle\") pod \"ovn-controller-l8fqd\" (UID: \"36eb99a8-609a-4af2-9fa7-d0051806659e\") " pod="openstack/ovn-controller-l8fqd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.609750 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/36eb99a8-609a-4af2-9fa7-d0051806659e-scripts\") pod \"ovn-controller-l8fqd\" (UID: \"36eb99a8-609a-4af2-9fa7-d0051806659e\") " pod="openstack/ovn-controller-l8fqd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.633778 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gv9lx\" (UniqueName: \"kubernetes.io/projected/36eb99a8-609a-4af2-9fa7-d0051806659e-kube-api-access-gv9lx\") pod \"ovn-controller-l8fqd\" (UID: \"36eb99a8-609a-4af2-9fa7-d0051806659e\") " pod="openstack/ovn-controller-l8fqd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.705645 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/948a10c7-fb61-4110-a08e-b2fd42cb842a-scripts\") pod \"ovn-controller-ovs-6htcd\" (UID: \"948a10c7-fb61-4110-a08e-b2fd42cb842a\") " pod="openstack/ovn-controller-ovs-6htcd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.705725 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/948a10c7-fb61-4110-a08e-b2fd42cb842a-var-run\") pod \"ovn-controller-ovs-6htcd\" (UID: \"948a10c7-fb61-4110-a08e-b2fd42cb842a\") " pod="openstack/ovn-controller-ovs-6htcd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.705746 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/948a10c7-fb61-4110-a08e-b2fd42cb842a-etc-ovs\") pod \"ovn-controller-ovs-6htcd\" (UID: \"948a10c7-fb61-4110-a08e-b2fd42cb842a\") " pod="openstack/ovn-controller-ovs-6htcd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.705858 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/948a10c7-fb61-4110-a08e-b2fd42cb842a-var-lib\") pod \"ovn-controller-ovs-6htcd\" (UID: \"948a10c7-fb61-4110-a08e-b2fd42cb842a\") " pod="openstack/ovn-controller-ovs-6htcd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.705978 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/948a10c7-fb61-4110-a08e-b2fd42cb842a-var-log\") pod \"ovn-controller-ovs-6htcd\" (UID: \"948a10c7-fb61-4110-a08e-b2fd42cb842a\") " pod="openstack/ovn-controller-ovs-6htcd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.706024 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5knh6\" (UniqueName: \"kubernetes.io/projected/948a10c7-fb61-4110-a08e-b2fd42cb842a-kube-api-access-5knh6\") pod \"ovn-controller-ovs-6htcd\" (UID: \"948a10c7-fb61-4110-a08e-b2fd42cb842a\") " pod="openstack/ovn-controller-ovs-6htcd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.706532 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/948a10c7-fb61-4110-a08e-b2fd42cb842a-var-run\") pod \"ovn-controller-ovs-6htcd\" (UID: \"948a10c7-fb61-4110-a08e-b2fd42cb842a\") " pod="openstack/ovn-controller-ovs-6htcd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.706648 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/948a10c7-fb61-4110-a08e-b2fd42cb842a-var-log\") pod \"ovn-controller-ovs-6htcd\" (UID: \"948a10c7-fb61-4110-a08e-b2fd42cb842a\") " pod="openstack/ovn-controller-ovs-6htcd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.706668 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/948a10c7-fb61-4110-a08e-b2fd42cb842a-etc-ovs\") pod \"ovn-controller-ovs-6htcd\" (UID: \"948a10c7-fb61-4110-a08e-b2fd42cb842a\") " pod="openstack/ovn-controller-ovs-6htcd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.706752 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/948a10c7-fb61-4110-a08e-b2fd42cb842a-var-lib\") pod \"ovn-controller-ovs-6htcd\" (UID: \"948a10c7-fb61-4110-a08e-b2fd42cb842a\") " pod="openstack/ovn-controller-ovs-6htcd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.708773 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/948a10c7-fb61-4110-a08e-b2fd42cb842a-scripts\") pod \"ovn-controller-ovs-6htcd\" (UID: \"948a10c7-fb61-4110-a08e-b2fd42cb842a\") " pod="openstack/ovn-controller-ovs-6htcd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.731524 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5knh6\" (UniqueName: \"kubernetes.io/projected/948a10c7-fb61-4110-a08e-b2fd42cb842a-kube-api-access-5knh6\") pod \"ovn-controller-ovs-6htcd\" (UID: \"948a10c7-fb61-4110-a08e-b2fd42cb842a\") " pod="openstack/ovn-controller-ovs-6htcd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.767964 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-l8fqd" Oct 01 09:50:47 crc kubenswrapper[4787]: I1001 09:50:47.789544 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-6htcd" Oct 01 09:50:48 crc kubenswrapper[4787]: I1001 09:50:48.041438 4787 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 09:50:48 crc kubenswrapper[4787]: I1001 09:50:48.348050 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7924a2ac-f411-42d6-8b23-f22faf58351a","Type":"ContainerStarted","Data":"d8a40d6b81c8741b3d179971147b2dd2d0628f3c452e153036beb898db0d290c"} Oct 01 09:50:50 crc kubenswrapper[4787]: I1001 09:50:50.422528 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 01 09:50:50 crc kubenswrapper[4787]: I1001 09:50:50.435689 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 01 09:50:50 crc kubenswrapper[4787]: I1001 09:50:50.441312 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 01 09:50:50 crc kubenswrapper[4787]: I1001 09:50:50.441631 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 01 09:50:50 crc kubenswrapper[4787]: I1001 09:50:50.441851 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 01 09:50:50 crc kubenswrapper[4787]: I1001 09:50:50.447526 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-tr85b" Oct 01 09:50:50 crc kubenswrapper[4787]: I1001 09:50:50.449334 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 01 09:50:50 crc kubenswrapper[4787]: I1001 09:50:50.560807 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/696f6d7f-812c-4d36-a799-397b83edf291-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"696f6d7f-812c-4d36-a799-397b83edf291\") " pod="openstack/ovsdbserver-sb-0" Oct 01 09:50:50 crc kubenswrapper[4787]: I1001 09:50:50.560883 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"696f6d7f-812c-4d36-a799-397b83edf291\") " pod="openstack/ovsdbserver-sb-0" Oct 01 09:50:50 crc kubenswrapper[4787]: I1001 09:50:50.560931 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/696f6d7f-812c-4d36-a799-397b83edf291-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"696f6d7f-812c-4d36-a799-397b83edf291\") " pod="openstack/ovsdbserver-sb-0" Oct 01 09:50:50 crc kubenswrapper[4787]: I1001 09:50:50.560962 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/696f6d7f-812c-4d36-a799-397b83edf291-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"696f6d7f-812c-4d36-a799-397b83edf291\") " pod="openstack/ovsdbserver-sb-0" Oct 01 09:50:50 crc kubenswrapper[4787]: I1001 09:50:50.561003 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/696f6d7f-812c-4d36-a799-397b83edf291-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"696f6d7f-812c-4d36-a799-397b83edf291\") " pod="openstack/ovsdbserver-sb-0" Oct 01 09:50:50 crc kubenswrapper[4787]: I1001 09:50:50.561044 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nndxk\" (UniqueName: \"kubernetes.io/projected/696f6d7f-812c-4d36-a799-397b83edf291-kube-api-access-nndxk\") pod \"ovsdbserver-sb-0\" (UID: \"696f6d7f-812c-4d36-a799-397b83edf291\") " pod="openstack/ovsdbserver-sb-0" Oct 01 09:50:50 crc kubenswrapper[4787]: I1001 09:50:50.561117 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/696f6d7f-812c-4d36-a799-397b83edf291-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"696f6d7f-812c-4d36-a799-397b83edf291\") " pod="openstack/ovsdbserver-sb-0" Oct 01 09:50:50 crc kubenswrapper[4787]: I1001 09:50:50.561144 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/696f6d7f-812c-4d36-a799-397b83edf291-config\") pod \"ovsdbserver-sb-0\" (UID: \"696f6d7f-812c-4d36-a799-397b83edf291\") " pod="openstack/ovsdbserver-sb-0" Oct 01 09:50:50 crc kubenswrapper[4787]: I1001 09:50:50.663536 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/696f6d7f-812c-4d36-a799-397b83edf291-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"696f6d7f-812c-4d36-a799-397b83edf291\") " pod="openstack/ovsdbserver-sb-0" Oct 01 09:50:50 crc kubenswrapper[4787]: I1001 09:50:50.663622 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/696f6d7f-812c-4d36-a799-397b83edf291-config\") pod \"ovsdbserver-sb-0\" (UID: \"696f6d7f-812c-4d36-a799-397b83edf291\") " pod="openstack/ovsdbserver-sb-0" Oct 01 09:50:50 crc kubenswrapper[4787]: I1001 09:50:50.663701 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/696f6d7f-812c-4d36-a799-397b83edf291-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"696f6d7f-812c-4d36-a799-397b83edf291\") " pod="openstack/ovsdbserver-sb-0" Oct 01 09:50:50 crc kubenswrapper[4787]: I1001 09:50:50.663740 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"696f6d7f-812c-4d36-a799-397b83edf291\") " pod="openstack/ovsdbserver-sb-0" Oct 01 09:50:50 crc kubenswrapper[4787]: I1001 09:50:50.663782 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/696f6d7f-812c-4d36-a799-397b83edf291-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"696f6d7f-812c-4d36-a799-397b83edf291\") " pod="openstack/ovsdbserver-sb-0" Oct 01 09:50:50 crc kubenswrapper[4787]: I1001 09:50:50.663813 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/696f6d7f-812c-4d36-a799-397b83edf291-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"696f6d7f-812c-4d36-a799-397b83edf291\") " pod="openstack/ovsdbserver-sb-0" Oct 01 09:50:50 crc kubenswrapper[4787]: I1001 09:50:50.663868 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/696f6d7f-812c-4d36-a799-397b83edf291-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"696f6d7f-812c-4d36-a799-397b83edf291\") " pod="openstack/ovsdbserver-sb-0" Oct 01 09:50:50 crc kubenswrapper[4787]: I1001 09:50:50.663928 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nndxk\" (UniqueName: \"kubernetes.io/projected/696f6d7f-812c-4d36-a799-397b83edf291-kube-api-access-nndxk\") pod \"ovsdbserver-sb-0\" (UID: \"696f6d7f-812c-4d36-a799-397b83edf291\") " pod="openstack/ovsdbserver-sb-0" Oct 01 09:50:50 crc kubenswrapper[4787]: I1001 09:50:50.665124 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/696f6d7f-812c-4d36-a799-397b83edf291-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"696f6d7f-812c-4d36-a799-397b83edf291\") " pod="openstack/ovsdbserver-sb-0" Oct 01 09:50:50 crc kubenswrapper[4787]: I1001 09:50:50.665243 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/696f6d7f-812c-4d36-a799-397b83edf291-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"696f6d7f-812c-4d36-a799-397b83edf291\") " pod="openstack/ovsdbserver-sb-0" Oct 01 09:50:50 crc kubenswrapper[4787]: I1001 09:50:50.665371 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"696f6d7f-812c-4d36-a799-397b83edf291\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/ovsdbserver-sb-0" Oct 01 09:50:50 crc kubenswrapper[4787]: I1001 09:50:50.667036 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/696f6d7f-812c-4d36-a799-397b83edf291-config\") pod \"ovsdbserver-sb-0\" (UID: \"696f6d7f-812c-4d36-a799-397b83edf291\") " pod="openstack/ovsdbserver-sb-0" Oct 01 09:50:50 crc kubenswrapper[4787]: I1001 09:50:50.668939 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/696f6d7f-812c-4d36-a799-397b83edf291-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"696f6d7f-812c-4d36-a799-397b83edf291\") " pod="openstack/ovsdbserver-sb-0" Oct 01 09:50:50 crc kubenswrapper[4787]: I1001 09:50:50.669158 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/696f6d7f-812c-4d36-a799-397b83edf291-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"696f6d7f-812c-4d36-a799-397b83edf291\") " pod="openstack/ovsdbserver-sb-0" Oct 01 09:50:50 crc kubenswrapper[4787]: I1001 09:50:50.677140 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/696f6d7f-812c-4d36-a799-397b83edf291-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"696f6d7f-812c-4d36-a799-397b83edf291\") " pod="openstack/ovsdbserver-sb-0" Oct 01 09:50:50 crc kubenswrapper[4787]: I1001 09:50:50.680314 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nndxk\" (UniqueName: \"kubernetes.io/projected/696f6d7f-812c-4d36-a799-397b83edf291-kube-api-access-nndxk\") pod \"ovsdbserver-sb-0\" (UID: \"696f6d7f-812c-4d36-a799-397b83edf291\") " pod="openstack/ovsdbserver-sb-0" Oct 01 09:50:50 crc kubenswrapper[4787]: I1001 09:50:50.711744 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"696f6d7f-812c-4d36-a799-397b83edf291\") " pod="openstack/ovsdbserver-sb-0" Oct 01 09:50:50 crc kubenswrapper[4787]: I1001 09:50:50.773088 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 01 09:50:53 crc kubenswrapper[4787]: E1001 09:50:53.729919 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:6276771339c90f342673dcaf7faa8c46e2c0ece62ed5efc4b7d65a095dabe07b" Oct 01 09:50:53 crc kubenswrapper[4787]: E1001 09:50:53.730658 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:6276771339c90f342673dcaf7faa8c46e2c0ece62ed5efc4b7d65a095dabe07b,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8dr9d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-b8b69cf79-6w6rh_openstack(0bdbc459-b8b9-4a6f-9462-61bd54c87b1d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 09:50:53 crc kubenswrapper[4787]: E1001 09:50:53.733164 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-b8b69cf79-6w6rh" podUID="0bdbc459-b8b9-4a6f-9462-61bd54c87b1d" Oct 01 09:50:53 crc kubenswrapper[4787]: I1001 09:50:53.940508 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 01 09:50:54 crc kubenswrapper[4787]: I1001 09:50:54.110492 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 01 09:50:55 crc kubenswrapper[4787]: I1001 09:50:55.336775 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8b69cf79-6w6rh" Oct 01 09:50:55 crc kubenswrapper[4787]: I1001 09:50:55.434657 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"7137d092-ff92-4b8d-84c5-9d9d269ddcc7","Type":"ContainerStarted","Data":"4bf1130c519a1f128f9d5f5a96e560922c2a9efbbc9a5c216d96b3c161972b8f"} Oct 01 09:50:55 crc kubenswrapper[4787]: I1001 09:50:55.436225 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c","Type":"ContainerStarted","Data":"4639df84bd0c8217d53d65103b1a5313d00bb3c863698707b6705a0cd4ad6842"} Oct 01 09:50:55 crc kubenswrapper[4787]: I1001 09:50:55.437966 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8b69cf79-6w6rh" event={"ID":"0bdbc459-b8b9-4a6f-9462-61bd54c87b1d","Type":"ContainerDied","Data":"9004515597def8ef11563c7909dedc26484b860ac226358d3eb15cb9c70bdf6d"} Oct 01 09:50:55 crc kubenswrapper[4787]: I1001 09:50:55.438102 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8b69cf79-6w6rh" Oct 01 09:50:55 crc kubenswrapper[4787]: I1001 09:50:55.489745 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bdbc459-b8b9-4a6f-9462-61bd54c87b1d-config\") pod \"0bdbc459-b8b9-4a6f-9462-61bd54c87b1d\" (UID: \"0bdbc459-b8b9-4a6f-9462-61bd54c87b1d\") " Oct 01 09:50:55 crc kubenswrapper[4787]: I1001 09:50:55.489847 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8dr9d\" (UniqueName: \"kubernetes.io/projected/0bdbc459-b8b9-4a6f-9462-61bd54c87b1d-kube-api-access-8dr9d\") pod \"0bdbc459-b8b9-4a6f-9462-61bd54c87b1d\" (UID: \"0bdbc459-b8b9-4a6f-9462-61bd54c87b1d\") " Oct 01 09:50:55 crc kubenswrapper[4787]: I1001 09:50:55.490858 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0bdbc459-b8b9-4a6f-9462-61bd54c87b1d-config" (OuterVolumeSpecName: "config") pod "0bdbc459-b8b9-4a6f-9462-61bd54c87b1d" (UID: "0bdbc459-b8b9-4a6f-9462-61bd54c87b1d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:50:55 crc kubenswrapper[4787]: I1001 09:50:55.499499 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bdbc459-b8b9-4a6f-9462-61bd54c87b1d-kube-api-access-8dr9d" (OuterVolumeSpecName: "kube-api-access-8dr9d") pod "0bdbc459-b8b9-4a6f-9462-61bd54c87b1d" (UID: "0bdbc459-b8b9-4a6f-9462-61bd54c87b1d"). InnerVolumeSpecName "kube-api-access-8dr9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:50:55 crc kubenswrapper[4787]: I1001 09:50:55.592068 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bdbc459-b8b9-4a6f-9462-61bd54c87b1d-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:50:55 crc kubenswrapper[4787]: I1001 09:50:55.592536 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8dr9d\" (UniqueName: \"kubernetes.io/projected/0bdbc459-b8b9-4a6f-9462-61bd54c87b1d-kube-api-access-8dr9d\") on node \"crc\" DevicePath \"\"" Oct 01 09:50:55 crc kubenswrapper[4787]: I1001 09:50:55.623058 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 09:50:55 crc kubenswrapper[4787]: I1001 09:50:55.811461 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8b69cf79-6w6rh"] Oct 01 09:50:55 crc kubenswrapper[4787]: I1001 09:50:55.826294 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8b69cf79-6w6rh"] Oct 01 09:50:55 crc kubenswrapper[4787]: I1001 09:50:55.908976 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 01 09:50:56 crc kubenswrapper[4787]: I1001 09:50:56.534843 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bdbc459-b8b9-4a6f-9462-61bd54c87b1d" path="/var/lib/kubelet/pods/0bdbc459-b8b9-4a6f-9462-61bd54c87b1d/volumes" Oct 01 09:50:56 crc kubenswrapper[4787]: I1001 09:50:56.863260 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-6htcd"] Oct 01 09:50:57 crc kubenswrapper[4787]: W1001 09:50:57.536372 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod469f4dc6_d176_4497_ac69_6130a375737b.slice/crio-1f63527c8f398eb15518a149dbb2c1123e7fe6b54914e56a50fa3f06f07cb2fa WatchSource:0}: Error finding container 1f63527c8f398eb15518a149dbb2c1123e7fe6b54914e56a50fa3f06f07cb2fa: Status 404 returned error can't find the container with id 1f63527c8f398eb15518a149dbb2c1123e7fe6b54914e56a50fa3f06f07cb2fa Oct 01 09:50:57 crc kubenswrapper[4787]: W1001 09:50:57.539099 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee50aeec_8875_45c2_80e2_7121480ae89f.slice/crio-fe36720e15abfd1f659416436cc2f03273337849cdafb44077f75a2d610b060f WatchSource:0}: Error finding container fe36720e15abfd1f659416436cc2f03273337849cdafb44077f75a2d610b060f: Status 404 returned error can't find the container with id fe36720e15abfd1f659416436cc2f03273337849cdafb44077f75a2d610b060f Oct 01 09:50:58 crc kubenswrapper[4787]: I1001 09:50:58.096596 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-l8fqd"] Oct 01 09:50:58 crc kubenswrapper[4787]: I1001 09:50:58.297162 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 01 09:50:58 crc kubenswrapper[4787]: W1001 09:50:58.460343 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod696f6d7f_812c_4d36_a799_397b83edf291.slice/crio-90f721c36b161f224c2d71055b696d992b109392d7ed27816b894f3930372269 WatchSource:0}: Error finding container 90f721c36b161f224c2d71055b696d992b109392d7ed27816b894f3930372269: Status 404 returned error can't find the container with id 90f721c36b161f224c2d71055b696d992b109392d7ed27816b894f3930372269 Oct 01 09:50:58 crc kubenswrapper[4787]: W1001 09:50:58.463769 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod36eb99a8_609a_4af2_9fa7_d0051806659e.slice/crio-a981004b3494c81912db0c37ad8dbfc3eb03eaae2bf71d9d54e2105d7489360c WatchSource:0}: Error finding container a981004b3494c81912db0c37ad8dbfc3eb03eaae2bf71d9d54e2105d7489360c: Status 404 returned error can't find the container with id a981004b3494c81912db0c37ad8dbfc3eb03eaae2bf71d9d54e2105d7489360c Oct 01 09:50:58 crc kubenswrapper[4787]: I1001 09:50:58.464095 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"469f4dc6-d176-4497-ac69-6130a375737b","Type":"ContainerStarted","Data":"1f63527c8f398eb15518a149dbb2c1123e7fe6b54914e56a50fa3f06f07cb2fa"} Oct 01 09:50:58 crc kubenswrapper[4787]: I1001 09:50:58.466561 4787 generic.go:334] "Generic (PLEG): container finished" podID="105b8088-a178-492b-aefd-20a5e6ad0dc8" containerID="ae936474381271c575e1a56f21612c7a626e640004485f0662ae8025a0ac686f" exitCode=0 Oct 01 09:50:58 crc kubenswrapper[4787]: I1001 09:50:58.466696 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77795d58f5-kh2gz" event={"ID":"105b8088-a178-492b-aefd-20a5e6ad0dc8","Type":"ContainerDied","Data":"ae936474381271c575e1a56f21612c7a626e640004485f0662ae8025a0ac686f"} Oct 01 09:50:58 crc kubenswrapper[4787]: I1001 09:50:58.473683 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6htcd" event={"ID":"948a10c7-fb61-4110-a08e-b2fd42cb842a","Type":"ContainerStarted","Data":"b4797f7eebbac90d901c982ca40ff04470dfb26b8c59601578756581d510fba1"} Oct 01 09:50:58 crc kubenswrapper[4787]: I1001 09:50:58.478624 4787 generic.go:334] "Generic (PLEG): container finished" podID="47914662-d598-4ccf-af68-6d2d3a2d75b3" containerID="9621c91125a85632abcdc6f5a3b65a8e431c287f82d3a90a346d1b8086bd0df5" exitCode=0 Oct 01 09:50:58 crc kubenswrapper[4787]: I1001 09:50:58.478840 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d5f6f49c7-jqk5n" event={"ID":"47914662-d598-4ccf-af68-6d2d3a2d75b3","Type":"ContainerDied","Data":"9621c91125a85632abcdc6f5a3b65a8e431c287f82d3a90a346d1b8086bd0df5"} Oct 01 09:50:58 crc kubenswrapper[4787]: I1001 09:50:58.483674 4787 generic.go:334] "Generic (PLEG): container finished" podID="a5c38d07-c1a7-46fd-bcce-1680eff35b1f" containerID="b229ccb2cd047f4fad206e7ffdcf9f2176c9e3ebceb99e4c6319c4ef06d573b7" exitCode=0 Oct 01 09:50:58 crc kubenswrapper[4787]: I1001 09:50:58.483760 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b6f94bdfc-jdsn9" event={"ID":"a5c38d07-c1a7-46fd-bcce-1680eff35b1f","Type":"ContainerDied","Data":"b229ccb2cd047f4fad206e7ffdcf9f2176c9e3ebceb99e4c6319c4ef06d573b7"} Oct 01 09:50:58 crc kubenswrapper[4787]: I1001 09:50:58.486736 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c","Type":"ContainerStarted","Data":"cea1676f6d8a0fe1fd1e3f1ffe91dbf44da436ea6f2d368f971398152f6803f6"} Oct 01 09:50:58 crc kubenswrapper[4787]: I1001 09:50:58.500532 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7924a2ac-f411-42d6-8b23-f22faf58351a","Type":"ContainerStarted","Data":"bb1b251c96f3ea437f5cfeaa65acd3c4677dbeae917908746420733d6829908b"} Oct 01 09:50:58 crc kubenswrapper[4787]: I1001 09:50:58.505804 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ee50aeec-8875-45c2-80e2-7121480ae89f","Type":"ContainerStarted","Data":"fe36720e15abfd1f659416436cc2f03273337849cdafb44077f75a2d610b060f"} Oct 01 09:50:59 crc kubenswrapper[4787]: I1001 09:50:59.517853 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ce587847-47c2-41de-95d1-c9f8ab88961e","Type":"ContainerStarted","Data":"e46335d7899902dbb16ab47d03f56851aae7b293726f79ce1a33fe0fda821f6e"} Oct 01 09:50:59 crc kubenswrapper[4787]: I1001 09:50:59.523831 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-l8fqd" event={"ID":"36eb99a8-609a-4af2-9fa7-d0051806659e","Type":"ContainerStarted","Data":"a981004b3494c81912db0c37ad8dbfc3eb03eaae2bf71d9d54e2105d7489360c"} Oct 01 09:50:59 crc kubenswrapper[4787]: I1001 09:50:59.526297 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e762572f-d5b9-462e-9953-7143f648c9ae","Type":"ContainerStarted","Data":"2c2684e2732e0784cacbce917a9aa23a444a150b54d2e11ce18ac1efd3f00bd0"} Oct 01 09:50:59 crc kubenswrapper[4787]: I1001 09:50:59.528679 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"696f6d7f-812c-4d36-a799-397b83edf291","Type":"ContainerStarted","Data":"90f721c36b161f224c2d71055b696d992b109392d7ed27816b894f3930372269"} Oct 01 09:51:01 crc kubenswrapper[4787]: I1001 09:51:01.482877 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d5f6f49c7-jqk5n" Oct 01 09:51:01 crc kubenswrapper[4787]: I1001 09:51:01.545783 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d5f6f49c7-jqk5n" event={"ID":"47914662-d598-4ccf-af68-6d2d3a2d75b3","Type":"ContainerDied","Data":"e1cecaaac1a7d6793e33a1bfe9714746666ee7d8e097b27a8b5dc79a7c37c358"} Oct 01 09:51:01 crc kubenswrapper[4787]: I1001 09:51:01.545860 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d5f6f49c7-jqk5n" Oct 01 09:51:01 crc kubenswrapper[4787]: I1001 09:51:01.545873 4787 scope.go:117] "RemoveContainer" containerID="9621c91125a85632abcdc6f5a3b65a8e431c287f82d3a90a346d1b8086bd0df5" Oct 01 09:51:01 crc kubenswrapper[4787]: I1001 09:51:01.617315 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lh64v\" (UniqueName: \"kubernetes.io/projected/47914662-d598-4ccf-af68-6d2d3a2d75b3-kube-api-access-lh64v\") pod \"47914662-d598-4ccf-af68-6d2d3a2d75b3\" (UID: \"47914662-d598-4ccf-af68-6d2d3a2d75b3\") " Oct 01 09:51:01 crc kubenswrapper[4787]: I1001 09:51:01.617588 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/47914662-d598-4ccf-af68-6d2d3a2d75b3-dns-svc\") pod \"47914662-d598-4ccf-af68-6d2d3a2d75b3\" (UID: \"47914662-d598-4ccf-af68-6d2d3a2d75b3\") " Oct 01 09:51:01 crc kubenswrapper[4787]: I1001 09:51:01.617688 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47914662-d598-4ccf-af68-6d2d3a2d75b3-config\") pod \"47914662-d598-4ccf-af68-6d2d3a2d75b3\" (UID: \"47914662-d598-4ccf-af68-6d2d3a2d75b3\") " Oct 01 09:51:01 crc kubenswrapper[4787]: I1001 09:51:01.628277 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47914662-d598-4ccf-af68-6d2d3a2d75b3-kube-api-access-lh64v" (OuterVolumeSpecName: "kube-api-access-lh64v") pod "47914662-d598-4ccf-af68-6d2d3a2d75b3" (UID: "47914662-d598-4ccf-af68-6d2d3a2d75b3"). InnerVolumeSpecName "kube-api-access-lh64v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:51:01 crc kubenswrapper[4787]: I1001 09:51:01.640797 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47914662-d598-4ccf-af68-6d2d3a2d75b3-config" (OuterVolumeSpecName: "config") pod "47914662-d598-4ccf-af68-6d2d3a2d75b3" (UID: "47914662-d598-4ccf-af68-6d2d3a2d75b3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:51:01 crc kubenswrapper[4787]: I1001 09:51:01.662940 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47914662-d598-4ccf-af68-6d2d3a2d75b3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "47914662-d598-4ccf-af68-6d2d3a2d75b3" (UID: "47914662-d598-4ccf-af68-6d2d3a2d75b3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:51:01 crc kubenswrapper[4787]: I1001 09:51:01.720682 4787 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/47914662-d598-4ccf-af68-6d2d3a2d75b3-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:01 crc kubenswrapper[4787]: I1001 09:51:01.720716 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47914662-d598-4ccf-af68-6d2d3a2d75b3-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:01 crc kubenswrapper[4787]: I1001 09:51:01.720732 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lh64v\" (UniqueName: \"kubernetes.io/projected/47914662-d598-4ccf-af68-6d2d3a2d75b3-kube-api-access-lh64v\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:01 crc kubenswrapper[4787]: I1001 09:51:01.901596 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d5f6f49c7-jqk5n"] Oct 01 09:51:01 crc kubenswrapper[4787]: I1001 09:51:01.908388 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-d5f6f49c7-jqk5n"] Oct 01 09:51:02 crc kubenswrapper[4787]: I1001 09:51:02.534051 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47914662-d598-4ccf-af68-6d2d3a2d75b3" path="/var/lib/kubelet/pods/47914662-d598-4ccf-af68-6d2d3a2d75b3/volumes" Oct 01 09:51:04 crc kubenswrapper[4787]: I1001 09:51:04.574202 4787 generic.go:334] "Generic (PLEG): container finished" podID="37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c" containerID="cea1676f6d8a0fe1fd1e3f1ffe91dbf44da436ea6f2d368f971398152f6803f6" exitCode=0 Oct 01 09:51:04 crc kubenswrapper[4787]: I1001 09:51:04.574425 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c","Type":"ContainerDied","Data":"cea1676f6d8a0fe1fd1e3f1ffe91dbf44da436ea6f2d368f971398152f6803f6"} Oct 01 09:51:04 crc kubenswrapper[4787]: I1001 09:51:04.579193 4787 generic.go:334] "Generic (PLEG): container finished" podID="7924a2ac-f411-42d6-8b23-f22faf58351a" containerID="bb1b251c96f3ea437f5cfeaa65acd3c4677dbeae917908746420733d6829908b" exitCode=0 Oct 01 09:51:04 crc kubenswrapper[4787]: I1001 09:51:04.579251 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7924a2ac-f411-42d6-8b23-f22faf58351a","Type":"ContainerDied","Data":"bb1b251c96f3ea437f5cfeaa65acd3c4677dbeae917908746420733d6829908b"} Oct 01 09:51:05 crc kubenswrapper[4787]: I1001 09:51:05.594392 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77795d58f5-kh2gz" event={"ID":"105b8088-a178-492b-aefd-20a5e6ad0dc8","Type":"ContainerStarted","Data":"fd066a586c077d34157337d39d10e12ad72554813b7b93ab4f4c3567d5ef86c7"} Oct 01 09:51:05 crc kubenswrapper[4787]: I1001 09:51:05.595530 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77795d58f5-kh2gz" Oct 01 09:51:05 crc kubenswrapper[4787]: I1001 09:51:05.598046 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b6f94bdfc-jdsn9" event={"ID":"a5c38d07-c1a7-46fd-bcce-1680eff35b1f","Type":"ContainerStarted","Data":"cd2b8a3131f2ae66c7d291432884367c98409a4ad95ae7542cc937eb2930b2d6"} Oct 01 09:51:05 crc kubenswrapper[4787]: I1001 09:51:05.598319 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b6f94bdfc-jdsn9" Oct 01 09:51:05 crc kubenswrapper[4787]: I1001 09:51:05.618685 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77795d58f5-kh2gz" podStartSLOduration=9.954582559 podStartE2EDuration="29.618618334s" podCreationTimestamp="2025-10-01 09:50:36 +0000 UTC" firstStartedPulling="2025-10-01 09:50:37.919324531 +0000 UTC m=+870.034468688" lastFinishedPulling="2025-10-01 09:50:57.583360306 +0000 UTC m=+889.698504463" observedRunningTime="2025-10-01 09:51:05.615803382 +0000 UTC m=+897.730947539" watchObservedRunningTime="2025-10-01 09:51:05.618618334 +0000 UTC m=+897.733762491" Oct 01 09:51:05 crc kubenswrapper[4787]: I1001 09:51:05.645032 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b6f94bdfc-jdsn9" podStartSLOduration=9.328614657 podStartE2EDuration="29.645008899s" podCreationTimestamp="2025-10-01 09:50:36 +0000 UTC" firstStartedPulling="2025-10-01 09:50:37.40338882 +0000 UTC m=+869.518532977" lastFinishedPulling="2025-10-01 09:50:57.719783052 +0000 UTC m=+889.834927219" observedRunningTime="2025-10-01 09:51:05.639149959 +0000 UTC m=+897.754294136" watchObservedRunningTime="2025-10-01 09:51:05.645008899 +0000 UTC m=+897.760153056" Oct 01 09:51:11 crc kubenswrapper[4787]: I1001 09:51:11.250626 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:51:11 crc kubenswrapper[4787]: I1001 09:51:11.251256 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:51:11 crc kubenswrapper[4787]: I1001 09:51:11.720392 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b6f94bdfc-jdsn9" Oct 01 09:51:12 crc kubenswrapper[4787]: I1001 09:51:12.353255 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77795d58f5-kh2gz" Oct 01 09:51:12 crc kubenswrapper[4787]: I1001 09:51:12.448638 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b6f94bdfc-jdsn9"] Oct 01 09:51:12 crc kubenswrapper[4787]: I1001 09:51:12.646463 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b6f94bdfc-jdsn9" podUID="a5c38d07-c1a7-46fd-bcce-1680eff35b1f" containerName="dnsmasq-dns" containerID="cri-o://cd2b8a3131f2ae66c7d291432884367c98409a4ad95ae7542cc937eb2930b2d6" gracePeriod=10 Oct 01 09:51:14 crc kubenswrapper[4787]: E1001 09:51:14.260229 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb" Oct 01 09:51:14 crc kubenswrapper[4787]: E1001 09:51:14.260720 4787 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb" Oct 01 09:51:14 crc kubenswrapper[4787]: E1001 09:51:14.261042 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-state-metrics,Image:registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb,Command:[],Args:[--resources=pods --namespaces=openstack],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http-metrics,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},ContainerPort{Name:telemetry,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wj8f6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/livez,Port:{0 8080 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod kube-state-metrics-0_openstack(469f4dc6-d176-4497-ac69-6130a375737b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 01 09:51:14 crc kubenswrapper[4787]: E1001 09:51:14.262341 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/kube-state-metrics-0" podUID="469f4dc6-d176-4497-ac69-6130a375737b" Oct 01 09:51:14 crc kubenswrapper[4787]: I1001 09:51:14.481478 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b6f94bdfc-jdsn9" Oct 01 09:51:14 crc kubenswrapper[4787]: I1001 09:51:14.572928 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5c38d07-c1a7-46fd-bcce-1680eff35b1f-config\") pod \"a5c38d07-c1a7-46fd-bcce-1680eff35b1f\" (UID: \"a5c38d07-c1a7-46fd-bcce-1680eff35b1f\") " Oct 01 09:51:14 crc kubenswrapper[4787]: I1001 09:51:14.573049 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdmbr\" (UniqueName: \"kubernetes.io/projected/a5c38d07-c1a7-46fd-bcce-1680eff35b1f-kube-api-access-fdmbr\") pod \"a5c38d07-c1a7-46fd-bcce-1680eff35b1f\" (UID: \"a5c38d07-c1a7-46fd-bcce-1680eff35b1f\") " Oct 01 09:51:14 crc kubenswrapper[4787]: I1001 09:51:14.573113 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a5c38d07-c1a7-46fd-bcce-1680eff35b1f-dns-svc\") pod \"a5c38d07-c1a7-46fd-bcce-1680eff35b1f\" (UID: \"a5c38d07-c1a7-46fd-bcce-1680eff35b1f\") " Oct 01 09:51:14 crc kubenswrapper[4787]: I1001 09:51:14.584906 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5c38d07-c1a7-46fd-bcce-1680eff35b1f-kube-api-access-fdmbr" (OuterVolumeSpecName: "kube-api-access-fdmbr") pod "a5c38d07-c1a7-46fd-bcce-1680eff35b1f" (UID: "a5c38d07-c1a7-46fd-bcce-1680eff35b1f"). InnerVolumeSpecName "kube-api-access-fdmbr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:51:14 crc kubenswrapper[4787]: I1001 09:51:14.672898 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5c38d07-c1a7-46fd-bcce-1680eff35b1f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a5c38d07-c1a7-46fd-bcce-1680eff35b1f" (UID: "a5c38d07-c1a7-46fd-bcce-1680eff35b1f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:51:14 crc kubenswrapper[4787]: I1001 09:51:14.673880 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5c38d07-c1a7-46fd-bcce-1680eff35b1f-config" (OuterVolumeSpecName: "config") pod "a5c38d07-c1a7-46fd-bcce-1680eff35b1f" (UID: "a5c38d07-c1a7-46fd-bcce-1680eff35b1f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:51:14 crc kubenswrapper[4787]: I1001 09:51:14.674972 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdmbr\" (UniqueName: \"kubernetes.io/projected/a5c38d07-c1a7-46fd-bcce-1680eff35b1f-kube-api-access-fdmbr\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:14 crc kubenswrapper[4787]: I1001 09:51:14.675001 4787 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a5c38d07-c1a7-46fd-bcce-1680eff35b1f-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:14 crc kubenswrapper[4787]: I1001 09:51:14.675014 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5c38d07-c1a7-46fd-bcce-1680eff35b1f-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:14 crc kubenswrapper[4787]: I1001 09:51:14.678539 4787 generic.go:334] "Generic (PLEG): container finished" podID="a5c38d07-c1a7-46fd-bcce-1680eff35b1f" containerID="cd2b8a3131f2ae66c7d291432884367c98409a4ad95ae7542cc937eb2930b2d6" exitCode=0 Oct 01 09:51:14 crc kubenswrapper[4787]: I1001 09:51:14.678596 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b6f94bdfc-jdsn9" Oct 01 09:51:14 crc kubenswrapper[4787]: I1001 09:51:14.678664 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b6f94bdfc-jdsn9" event={"ID":"a5c38d07-c1a7-46fd-bcce-1680eff35b1f","Type":"ContainerDied","Data":"cd2b8a3131f2ae66c7d291432884367c98409a4ad95ae7542cc937eb2930b2d6"} Oct 01 09:51:14 crc kubenswrapper[4787]: I1001 09:51:14.678727 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b6f94bdfc-jdsn9" event={"ID":"a5c38d07-c1a7-46fd-bcce-1680eff35b1f","Type":"ContainerDied","Data":"b601caa8d6560d005c9b5fd5d227856c95031c41c0f73e2e9ef3ac890b073f64"} Oct 01 09:51:14 crc kubenswrapper[4787]: I1001 09:51:14.678751 4787 scope.go:117] "RemoveContainer" containerID="cd2b8a3131f2ae66c7d291432884367c98409a4ad95ae7542cc937eb2930b2d6" Oct 01 09:51:14 crc kubenswrapper[4787]: I1001 09:51:14.684675 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7924a2ac-f411-42d6-8b23-f22faf58351a","Type":"ContainerStarted","Data":"e94abaafad0b07a11dd703d5f6fdae8536e9219bff347d4e7348f4b0b1977bda"} Oct 01 09:51:14 crc kubenswrapper[4787]: I1001 09:51:14.692921 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ee50aeec-8875-45c2-80e2-7121480ae89f","Type":"ContainerStarted","Data":"c6c4c67f0d5f82c1d16526609f19c7ad0fe5ca4b9064ecd9727aaab0bdda0e41"} Oct 01 09:51:14 crc kubenswrapper[4787]: I1001 09:51:14.698527 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"696f6d7f-812c-4d36-a799-397b83edf291","Type":"ContainerStarted","Data":"9d2b6eb7915956a224bd608e9c66372530d8ceb9b75f76e89ee9c881134842b1"} Oct 01 09:51:14 crc kubenswrapper[4787]: I1001 09:51:14.702120 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6htcd" event={"ID":"948a10c7-fb61-4110-a08e-b2fd42cb842a","Type":"ContainerStarted","Data":"9791093e5783f8ebbb3a4ad03213fb41cd077009ee97e906d73b5b58272559e9"} Oct 01 09:51:14 crc kubenswrapper[4787]: I1001 09:51:14.705006 4787 scope.go:117] "RemoveContainer" containerID="b229ccb2cd047f4fad206e7ffdcf9f2176c9e3ebceb99e4c6319c4ef06d573b7" Oct 01 09:51:14 crc kubenswrapper[4787]: I1001 09:51:14.715282 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=26.079979161 podStartE2EDuration="35.715256031s" podCreationTimestamp="2025-10-01 09:50:39 +0000 UTC" firstStartedPulling="2025-10-01 09:50:48.041143051 +0000 UTC m=+880.156287208" lastFinishedPulling="2025-10-01 09:50:57.676419921 +0000 UTC m=+889.791564078" observedRunningTime="2025-10-01 09:51:14.708223201 +0000 UTC m=+906.823367358" watchObservedRunningTime="2025-10-01 09:51:14.715256031 +0000 UTC m=+906.830400188" Oct 01 09:51:14 crc kubenswrapper[4787]: I1001 09:51:14.719972 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"7137d092-ff92-4b8d-84c5-9d9d269ddcc7","Type":"ContainerStarted","Data":"59de3d8b747ebe0bbac65c188d3a854d18001ffaac43fff9c10a690751146831"} Oct 01 09:51:14 crc kubenswrapper[4787]: I1001 09:51:14.720180 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 01 09:51:14 crc kubenswrapper[4787]: I1001 09:51:14.723016 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-l8fqd" Oct 01 09:51:14 crc kubenswrapper[4787]: I1001 09:51:14.736965 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c","Type":"ContainerStarted","Data":"0358c8d669acd4e32de475bdaeb4ddda13fab3428d4a43a16588504a7b3e1ac9"} Oct 01 09:51:14 crc kubenswrapper[4787]: I1001 09:51:14.747815 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b6f94bdfc-jdsn9"] Oct 01 09:51:14 crc kubenswrapper[4787]: E1001 09:51:14.750380 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb\\\"\"" pod="openstack/kube-state-metrics-0" podUID="469f4dc6-d176-4497-ac69-6130a375737b" Oct 01 09:51:14 crc kubenswrapper[4787]: I1001 09:51:14.750583 4787 scope.go:117] "RemoveContainer" containerID="cd2b8a3131f2ae66c7d291432884367c98409a4ad95ae7542cc937eb2930b2d6" Oct 01 09:51:14 crc kubenswrapper[4787]: I1001 09:51:14.753409 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b6f94bdfc-jdsn9"] Oct 01 09:51:14 crc kubenswrapper[4787]: E1001 09:51:14.753473 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd2b8a3131f2ae66c7d291432884367c98409a4ad95ae7542cc937eb2930b2d6\": container with ID starting with cd2b8a3131f2ae66c7d291432884367c98409a4ad95ae7542cc937eb2930b2d6 not found: ID does not exist" containerID="cd2b8a3131f2ae66c7d291432884367c98409a4ad95ae7542cc937eb2930b2d6" Oct 01 09:51:14 crc kubenswrapper[4787]: I1001 09:51:14.753502 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd2b8a3131f2ae66c7d291432884367c98409a4ad95ae7542cc937eb2930b2d6"} err="failed to get container status \"cd2b8a3131f2ae66c7d291432884367c98409a4ad95ae7542cc937eb2930b2d6\": rpc error: code = NotFound desc = could not find container \"cd2b8a3131f2ae66c7d291432884367c98409a4ad95ae7542cc937eb2930b2d6\": container with ID starting with cd2b8a3131f2ae66c7d291432884367c98409a4ad95ae7542cc937eb2930b2d6 not found: ID does not exist" Oct 01 09:51:14 crc kubenswrapper[4787]: I1001 09:51:14.753521 4787 scope.go:117] "RemoveContainer" containerID="b229ccb2cd047f4fad206e7ffdcf9f2176c9e3ebceb99e4c6319c4ef06d573b7" Oct 01 09:51:14 crc kubenswrapper[4787]: E1001 09:51:14.753896 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b229ccb2cd047f4fad206e7ffdcf9f2176c9e3ebceb99e4c6319c4ef06d573b7\": container with ID starting with b229ccb2cd047f4fad206e7ffdcf9f2176c9e3ebceb99e4c6319c4ef06d573b7 not found: ID does not exist" containerID="b229ccb2cd047f4fad206e7ffdcf9f2176c9e3ebceb99e4c6319c4ef06d573b7" Oct 01 09:51:14 crc kubenswrapper[4787]: I1001 09:51:14.753933 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b229ccb2cd047f4fad206e7ffdcf9f2176c9e3ebceb99e4c6319c4ef06d573b7"} err="failed to get container status \"b229ccb2cd047f4fad206e7ffdcf9f2176c9e3ebceb99e4c6319c4ef06d573b7\": rpc error: code = NotFound desc = could not find container \"b229ccb2cd047f4fad206e7ffdcf9f2176c9e3ebceb99e4c6319c4ef06d573b7\": container with ID starting with b229ccb2cd047f4fad206e7ffdcf9f2176c9e3ebceb99e4c6319c4ef06d573b7 not found: ID does not exist" Oct 01 09:51:14 crc kubenswrapper[4787]: I1001 09:51:14.771345 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=34.299001801 podStartE2EDuration="36.771319728s" podCreationTimestamp="2025-10-01 09:50:38 +0000 UTC" firstStartedPulling="2025-10-01 09:50:55.245594207 +0000 UTC m=+887.360738364" lastFinishedPulling="2025-10-01 09:50:57.717912134 +0000 UTC m=+889.833056291" observedRunningTime="2025-10-01 09:51:14.764935184 +0000 UTC m=+906.880079371" watchObservedRunningTime="2025-10-01 09:51:14.771319728 +0000 UTC m=+906.886463885" Oct 01 09:51:14 crc kubenswrapper[4787]: I1001 09:51:14.795194 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-l8fqd" podStartSLOduration=21.511176931 podStartE2EDuration="27.795176719s" podCreationTimestamp="2025-10-01 09:50:47 +0000 UTC" firstStartedPulling="2025-10-01 09:50:58.467616737 +0000 UTC m=+890.582760894" lastFinishedPulling="2025-10-01 09:51:04.751616525 +0000 UTC m=+896.866760682" observedRunningTime="2025-10-01 09:51:14.790764045 +0000 UTC m=+906.905908222" watchObservedRunningTime="2025-10-01 09:51:14.795176719 +0000 UTC m=+906.910320876" Oct 01 09:51:14 crc kubenswrapper[4787]: I1001 09:51:14.812292 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=26.886165647 podStartE2EDuration="34.812269447s" podCreationTimestamp="2025-10-01 09:50:40 +0000 UTC" firstStartedPulling="2025-10-01 09:50:55.244721864 +0000 UTC m=+887.359866021" lastFinishedPulling="2025-10-01 09:51:03.170825664 +0000 UTC m=+895.285969821" observedRunningTime="2025-10-01 09:51:14.808422648 +0000 UTC m=+906.923566815" watchObservedRunningTime="2025-10-01 09:51:14.812269447 +0000 UTC m=+906.927413614" Oct 01 09:51:15 crc kubenswrapper[4787]: I1001 09:51:15.760730 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-l8fqd" event={"ID":"36eb99a8-609a-4af2-9fa7-d0051806659e","Type":"ContainerStarted","Data":"131693cf91aa44d36a6470514448e6b75ac85bdf904f35ad696c790c23982ed2"} Oct 01 09:51:15 crc kubenswrapper[4787]: I1001 09:51:15.768594 4787 generic.go:334] "Generic (PLEG): container finished" podID="948a10c7-fb61-4110-a08e-b2fd42cb842a" containerID="9791093e5783f8ebbb3a4ad03213fb41cd077009ee97e906d73b5b58272559e9" exitCode=0 Oct 01 09:51:15 crc kubenswrapper[4787]: I1001 09:51:15.768681 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6htcd" event={"ID":"948a10c7-fb61-4110-a08e-b2fd42cb842a","Type":"ContainerDied","Data":"9791093e5783f8ebbb3a4ad03213fb41cd077009ee97e906d73b5b58272559e9"} Oct 01 09:51:16 crc kubenswrapper[4787]: E1001 09:51:16.133008 4787 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.129.56.39:47604->38.129.56.39:39781: write tcp 38.129.56.39:47604->38.129.56.39:39781: write: broken pipe Oct 01 09:51:16 crc kubenswrapper[4787]: I1001 09:51:16.540830 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5c38d07-c1a7-46fd-bcce-1680eff35b1f" path="/var/lib/kubelet/pods/a5c38d07-c1a7-46fd-bcce-1680eff35b1f/volumes" Oct 01 09:51:16 crc kubenswrapper[4787]: I1001 09:51:16.782086 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6htcd" event={"ID":"948a10c7-fb61-4110-a08e-b2fd42cb842a","Type":"ContainerStarted","Data":"793a44f0c9a9c91ffef0ef5f656ca4c80a200f01f8a0e7af7153315ad9b05f12"} Oct 01 09:51:18 crc kubenswrapper[4787]: I1001 09:51:18.800151 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ee50aeec-8875-45c2-80e2-7121480ae89f","Type":"ContainerStarted","Data":"348322c2c9f50b4cc3f66377e7e376b49e3e66a486671f6a21035b96d9765b61"} Oct 01 09:51:18 crc kubenswrapper[4787]: I1001 09:51:18.804156 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"696f6d7f-812c-4d36-a799-397b83edf291","Type":"ContainerStarted","Data":"4502084b0f92ba7c177132957e688066ae22119d773c2a7ccdbdc663cbfed0b8"} Oct 01 09:51:18 crc kubenswrapper[4787]: I1001 09:51:18.808572 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6htcd" event={"ID":"948a10c7-fb61-4110-a08e-b2fd42cb842a","Type":"ContainerStarted","Data":"0d83290f73f2b664f756fc6f6f6df421984f8a3d2875b985e1179aeb1bd09315"} Oct 01 09:51:18 crc kubenswrapper[4787]: I1001 09:51:18.808946 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-6htcd" Oct 01 09:51:18 crc kubenswrapper[4787]: I1001 09:51:18.808976 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-6htcd" Oct 01 09:51:18 crc kubenswrapper[4787]: I1001 09:51:18.825144 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=13.818785397 podStartE2EDuration="33.825125683s" podCreationTimestamp="2025-10-01 09:50:45 +0000 UTC" firstStartedPulling="2025-10-01 09:50:57.542613962 +0000 UTC m=+889.657758119" lastFinishedPulling="2025-10-01 09:51:17.548954248 +0000 UTC m=+909.664098405" observedRunningTime="2025-10-01 09:51:18.819108359 +0000 UTC m=+910.934252526" watchObservedRunningTime="2025-10-01 09:51:18.825125683 +0000 UTC m=+910.940269840" Oct 01 09:51:18 crc kubenswrapper[4787]: I1001 09:51:18.852729 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-6htcd" podStartSLOduration=25.339085566 podStartE2EDuration="31.85271323s" podCreationTimestamp="2025-10-01 09:50:47 +0000 UTC" firstStartedPulling="2025-10-01 09:50:57.597031946 +0000 UTC m=+889.712176103" lastFinishedPulling="2025-10-01 09:51:04.11065961 +0000 UTC m=+896.225803767" observedRunningTime="2025-10-01 09:51:18.843939865 +0000 UTC m=+910.959084052" watchObservedRunningTime="2025-10-01 09:51:18.85271323 +0000 UTC m=+910.967857387" Oct 01 09:51:18 crc kubenswrapper[4787]: I1001 09:51:18.871475 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=10.769600442 podStartE2EDuration="29.87145639s" podCreationTimestamp="2025-10-01 09:50:49 +0000 UTC" firstStartedPulling="2025-10-01 09:50:58.463697026 +0000 UTC m=+890.578841193" lastFinishedPulling="2025-10-01 09:51:17.565552984 +0000 UTC m=+909.680697141" observedRunningTime="2025-10-01 09:51:18.86673719 +0000 UTC m=+910.981881357" watchObservedRunningTime="2025-10-01 09:51:18.87145639 +0000 UTC m=+910.986600557" Oct 01 09:51:19 crc kubenswrapper[4787]: I1001 09:51:19.800601 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 01 09:51:19 crc kubenswrapper[4787]: I1001 09:51:19.800775 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 01 09:51:19 crc kubenswrapper[4787]: I1001 09:51:19.866825 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 01 09:51:20 crc kubenswrapper[4787]: I1001 09:51:20.187520 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 01 09:51:20 crc kubenswrapper[4787]: I1001 09:51:20.225672 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 01 09:51:20 crc kubenswrapper[4787]: I1001 09:51:20.773286 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 01 09:51:20 crc kubenswrapper[4787]: I1001 09:51:20.773324 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 01 09:51:20 crc kubenswrapper[4787]: I1001 09:51:20.811578 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 01 09:51:20 crc kubenswrapper[4787]: I1001 09:51:20.824775 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 01 09:51:20 crc kubenswrapper[4787]: I1001 09:51:20.861821 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 01 09:51:20 crc kubenswrapper[4787]: I1001 09:51:20.872007 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 01 09:51:20 crc kubenswrapper[4787]: I1001 09:51:20.872296 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 01 09:51:20 crc kubenswrapper[4787]: I1001 09:51:20.888911 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 01 09:51:20 crc kubenswrapper[4787]: I1001 09:51:20.889263 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 01 09:51:20 crc kubenswrapper[4787]: I1001 09:51:20.962041 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.084386 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86f865fc45-784xp"] Oct 01 09:51:21 crc kubenswrapper[4787]: E1001 09:51:21.084734 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47914662-d598-4ccf-af68-6d2d3a2d75b3" containerName="init" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.084753 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="47914662-d598-4ccf-af68-6d2d3a2d75b3" containerName="init" Oct 01 09:51:21 crc kubenswrapper[4787]: E1001 09:51:21.084764 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5c38d07-c1a7-46fd-bcce-1680eff35b1f" containerName="dnsmasq-dns" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.084771 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5c38d07-c1a7-46fd-bcce-1680eff35b1f" containerName="dnsmasq-dns" Oct 01 09:51:21 crc kubenswrapper[4787]: E1001 09:51:21.084797 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5c38d07-c1a7-46fd-bcce-1680eff35b1f" containerName="init" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.084803 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5c38d07-c1a7-46fd-bcce-1680eff35b1f" containerName="init" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.084966 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="47914662-d598-4ccf-af68-6d2d3a2d75b3" containerName="init" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.084984 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5c38d07-c1a7-46fd-bcce-1680eff35b1f" containerName="dnsmasq-dns" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.085791 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86f865fc45-784xp" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.087487 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.100854 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86f865fc45-784xp"] Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.196093 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-d9cxp"] Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.197397 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-d9cxp" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.205330 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.211925 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/23cec814-a12e-4b66-922a-ea061ed7e10a-ovsdbserver-sb\") pod \"dnsmasq-dns-86f865fc45-784xp\" (UID: \"23cec814-a12e-4b66-922a-ea061ed7e10a\") " pod="openstack/dnsmasq-dns-86f865fc45-784xp" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.211980 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23cec814-a12e-4b66-922a-ea061ed7e10a-config\") pod \"dnsmasq-dns-86f865fc45-784xp\" (UID: \"23cec814-a12e-4b66-922a-ea061ed7e10a\") " pod="openstack/dnsmasq-dns-86f865fc45-784xp" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.212004 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/23cec814-a12e-4b66-922a-ea061ed7e10a-dns-svc\") pod \"dnsmasq-dns-86f865fc45-784xp\" (UID: \"23cec814-a12e-4b66-922a-ea061ed7e10a\") " pod="openstack/dnsmasq-dns-86f865fc45-784xp" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.212096 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2t2f\" (UniqueName: \"kubernetes.io/projected/23cec814-a12e-4b66-922a-ea061ed7e10a-kube-api-access-s2t2f\") pod \"dnsmasq-dns-86f865fc45-784xp\" (UID: \"23cec814-a12e-4b66-922a-ea061ed7e10a\") " pod="openstack/dnsmasq-dns-86f865fc45-784xp" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.212836 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-d9cxp"] Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.235773 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.258025 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-4lszz"] Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.259148 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-4lszz" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.266533 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-4lszz"] Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.313772 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/23cec814-a12e-4b66-922a-ea061ed7e10a-ovsdbserver-sb\") pod \"dnsmasq-dns-86f865fc45-784xp\" (UID: \"23cec814-a12e-4b66-922a-ea061ed7e10a\") " pod="openstack/dnsmasq-dns-86f865fc45-784xp" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.313833 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11dbf4b0-9694-40e2-a122-9da928c50ad8-config\") pod \"ovn-controller-metrics-d9cxp\" (UID: \"11dbf4b0-9694-40e2-a122-9da928c50ad8\") " pod="openstack/ovn-controller-metrics-d9cxp" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.313862 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/11dbf4b0-9694-40e2-a122-9da928c50ad8-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-d9cxp\" (UID: \"11dbf4b0-9694-40e2-a122-9da928c50ad8\") " pod="openstack/ovn-controller-metrics-d9cxp" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.313900 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nd7pk\" (UniqueName: \"kubernetes.io/projected/11413a2c-48bf-4160-8e54-d843305b3b97-kube-api-access-nd7pk\") pod \"placement-db-create-4lszz\" (UID: \"11413a2c-48bf-4160-8e54-d843305b3b97\") " pod="openstack/placement-db-create-4lszz" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.313925 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23cec814-a12e-4b66-922a-ea061ed7e10a-config\") pod \"dnsmasq-dns-86f865fc45-784xp\" (UID: \"23cec814-a12e-4b66-922a-ea061ed7e10a\") " pod="openstack/dnsmasq-dns-86f865fc45-784xp" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.313953 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/11dbf4b0-9694-40e2-a122-9da928c50ad8-ovs-rundir\") pod \"ovn-controller-metrics-d9cxp\" (UID: \"11dbf4b0-9694-40e2-a122-9da928c50ad8\") " pod="openstack/ovn-controller-metrics-d9cxp" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.313976 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11dbf4b0-9694-40e2-a122-9da928c50ad8-combined-ca-bundle\") pod \"ovn-controller-metrics-d9cxp\" (UID: \"11dbf4b0-9694-40e2-a122-9da928c50ad8\") " pod="openstack/ovn-controller-metrics-d9cxp" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.314002 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/23cec814-a12e-4b66-922a-ea061ed7e10a-dns-svc\") pod \"dnsmasq-dns-86f865fc45-784xp\" (UID: \"23cec814-a12e-4b66-922a-ea061ed7e10a\") " pod="openstack/dnsmasq-dns-86f865fc45-784xp" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.314039 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/11dbf4b0-9694-40e2-a122-9da928c50ad8-ovn-rundir\") pod \"ovn-controller-metrics-d9cxp\" (UID: \"11dbf4b0-9694-40e2-a122-9da928c50ad8\") " pod="openstack/ovn-controller-metrics-d9cxp" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.314237 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbm48\" (UniqueName: \"kubernetes.io/projected/11dbf4b0-9694-40e2-a122-9da928c50ad8-kube-api-access-nbm48\") pod \"ovn-controller-metrics-d9cxp\" (UID: \"11dbf4b0-9694-40e2-a122-9da928c50ad8\") " pod="openstack/ovn-controller-metrics-d9cxp" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.314414 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2t2f\" (UniqueName: \"kubernetes.io/projected/23cec814-a12e-4b66-922a-ea061ed7e10a-kube-api-access-s2t2f\") pod \"dnsmasq-dns-86f865fc45-784xp\" (UID: \"23cec814-a12e-4b66-922a-ea061ed7e10a\") " pod="openstack/dnsmasq-dns-86f865fc45-784xp" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.315555 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23cec814-a12e-4b66-922a-ea061ed7e10a-config\") pod \"dnsmasq-dns-86f865fc45-784xp\" (UID: \"23cec814-a12e-4b66-922a-ea061ed7e10a\") " pod="openstack/dnsmasq-dns-86f865fc45-784xp" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.316053 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/23cec814-a12e-4b66-922a-ea061ed7e10a-ovsdbserver-sb\") pod \"dnsmasq-dns-86f865fc45-784xp\" (UID: \"23cec814-a12e-4b66-922a-ea061ed7e10a\") " pod="openstack/dnsmasq-dns-86f865fc45-784xp" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.319905 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/23cec814-a12e-4b66-922a-ea061ed7e10a-dns-svc\") pod \"dnsmasq-dns-86f865fc45-784xp\" (UID: \"23cec814-a12e-4b66-922a-ea061ed7e10a\") " pod="openstack/dnsmasq-dns-86f865fc45-784xp" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.327054 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86f865fc45-784xp"] Oct 01 09:51:21 crc kubenswrapper[4787]: E1001 09:51:21.328591 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-s2t2f], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-86f865fc45-784xp" podUID="23cec814-a12e-4b66-922a-ea061ed7e10a" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.338911 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2t2f\" (UniqueName: \"kubernetes.io/projected/23cec814-a12e-4b66-922a-ea061ed7e10a-kube-api-access-s2t2f\") pod \"dnsmasq-dns-86f865fc45-784xp\" (UID: \"23cec814-a12e-4b66-922a-ea061ed7e10a\") " pod="openstack/dnsmasq-dns-86f865fc45-784xp" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.363060 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d86d68bf7-mn47f"] Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.365127 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d86d68bf7-mn47f" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.374206 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.405220 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d86d68bf7-mn47f"] Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.417195 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/11dbf4b0-9694-40e2-a122-9da928c50ad8-ovn-rundir\") pod \"ovn-controller-metrics-d9cxp\" (UID: \"11dbf4b0-9694-40e2-a122-9da928c50ad8\") " pod="openstack/ovn-controller-metrics-d9cxp" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.417592 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26ae359c-09e5-48c8-8f17-a5e6d7b75529-ovsdbserver-sb\") pod \"dnsmasq-dns-5d86d68bf7-mn47f\" (UID: \"26ae359c-09e5-48c8-8f17-a5e6d7b75529\") " pod="openstack/dnsmasq-dns-5d86d68bf7-mn47f" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.417649 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26ae359c-09e5-48c8-8f17-a5e6d7b75529-dns-svc\") pod \"dnsmasq-dns-5d86d68bf7-mn47f\" (UID: \"26ae359c-09e5-48c8-8f17-a5e6d7b75529\") " pod="openstack/dnsmasq-dns-5d86d68bf7-mn47f" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.417673 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbm48\" (UniqueName: \"kubernetes.io/projected/11dbf4b0-9694-40e2-a122-9da928c50ad8-kube-api-access-nbm48\") pod \"ovn-controller-metrics-d9cxp\" (UID: \"11dbf4b0-9694-40e2-a122-9da928c50ad8\") " pod="openstack/ovn-controller-metrics-d9cxp" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.417736 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26ae359c-09e5-48c8-8f17-a5e6d7b75529-config\") pod \"dnsmasq-dns-5d86d68bf7-mn47f\" (UID: \"26ae359c-09e5-48c8-8f17-a5e6d7b75529\") " pod="openstack/dnsmasq-dns-5d86d68bf7-mn47f" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.417784 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11dbf4b0-9694-40e2-a122-9da928c50ad8-config\") pod \"ovn-controller-metrics-d9cxp\" (UID: \"11dbf4b0-9694-40e2-a122-9da928c50ad8\") " pod="openstack/ovn-controller-metrics-d9cxp" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.417807 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/11dbf4b0-9694-40e2-a122-9da928c50ad8-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-d9cxp\" (UID: \"11dbf4b0-9694-40e2-a122-9da928c50ad8\") " pod="openstack/ovn-controller-metrics-d9cxp" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.417841 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nd7pk\" (UniqueName: \"kubernetes.io/projected/11413a2c-48bf-4160-8e54-d843305b3b97-kube-api-access-nd7pk\") pod \"placement-db-create-4lszz\" (UID: \"11413a2c-48bf-4160-8e54-d843305b3b97\") " pod="openstack/placement-db-create-4lszz" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.417862 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26ae359c-09e5-48c8-8f17-a5e6d7b75529-ovsdbserver-nb\") pod \"dnsmasq-dns-5d86d68bf7-mn47f\" (UID: \"26ae359c-09e5-48c8-8f17-a5e6d7b75529\") " pod="openstack/dnsmasq-dns-5d86d68bf7-mn47f" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.417882 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87xzb\" (UniqueName: \"kubernetes.io/projected/26ae359c-09e5-48c8-8f17-a5e6d7b75529-kube-api-access-87xzb\") pod \"dnsmasq-dns-5d86d68bf7-mn47f\" (UID: \"26ae359c-09e5-48c8-8f17-a5e6d7b75529\") " pod="openstack/dnsmasq-dns-5d86d68bf7-mn47f" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.417903 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/11dbf4b0-9694-40e2-a122-9da928c50ad8-ovs-rundir\") pod \"ovn-controller-metrics-d9cxp\" (UID: \"11dbf4b0-9694-40e2-a122-9da928c50ad8\") " pod="openstack/ovn-controller-metrics-d9cxp" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.417928 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11dbf4b0-9694-40e2-a122-9da928c50ad8-combined-ca-bundle\") pod \"ovn-controller-metrics-d9cxp\" (UID: \"11dbf4b0-9694-40e2-a122-9da928c50ad8\") " pod="openstack/ovn-controller-metrics-d9cxp" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.423403 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11dbf4b0-9694-40e2-a122-9da928c50ad8-config\") pod \"ovn-controller-metrics-d9cxp\" (UID: \"11dbf4b0-9694-40e2-a122-9da928c50ad8\") " pod="openstack/ovn-controller-metrics-d9cxp" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.423662 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/11dbf4b0-9694-40e2-a122-9da928c50ad8-ovn-rundir\") pod \"ovn-controller-metrics-d9cxp\" (UID: \"11dbf4b0-9694-40e2-a122-9da928c50ad8\") " pod="openstack/ovn-controller-metrics-d9cxp" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.424755 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/11dbf4b0-9694-40e2-a122-9da928c50ad8-ovs-rundir\") pod \"ovn-controller-metrics-d9cxp\" (UID: \"11dbf4b0-9694-40e2-a122-9da928c50ad8\") " pod="openstack/ovn-controller-metrics-d9cxp" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.458526 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbm48\" (UniqueName: \"kubernetes.io/projected/11dbf4b0-9694-40e2-a122-9da928c50ad8-kube-api-access-nbm48\") pod \"ovn-controller-metrics-d9cxp\" (UID: \"11dbf4b0-9694-40e2-a122-9da928c50ad8\") " pod="openstack/ovn-controller-metrics-d9cxp" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.464690 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/11dbf4b0-9694-40e2-a122-9da928c50ad8-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-d9cxp\" (UID: \"11dbf4b0-9694-40e2-a122-9da928c50ad8\") " pod="openstack/ovn-controller-metrics-d9cxp" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.468494 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nd7pk\" (UniqueName: \"kubernetes.io/projected/11413a2c-48bf-4160-8e54-d843305b3b97-kube-api-access-nd7pk\") pod \"placement-db-create-4lszz\" (UID: \"11413a2c-48bf-4160-8e54-d843305b3b97\") " pod="openstack/placement-db-create-4lszz" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.470194 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.473937 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.474825 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11dbf4b0-9694-40e2-a122-9da928c50ad8-combined-ca-bundle\") pod \"ovn-controller-metrics-d9cxp\" (UID: \"11dbf4b0-9694-40e2-a122-9da928c50ad8\") " pod="openstack/ovn-controller-metrics-d9cxp" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.479007 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.479488 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-8jqm6" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.479630 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.479771 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.482385 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-67phv"] Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.488403 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-67phv" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.500739 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.519207 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-d9cxp" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.520352 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26ae359c-09e5-48c8-8f17-a5e6d7b75529-config\") pod \"dnsmasq-dns-5d86d68bf7-mn47f\" (UID: \"26ae359c-09e5-48c8-8f17-a5e6d7b75529\") " pod="openstack/dnsmasq-dns-5d86d68bf7-mn47f" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.520394 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wgtg\" (UniqueName: \"kubernetes.io/projected/29fb6825-d8d1-4401-b0ed-e1f3538cf4c5-kube-api-access-9wgtg\") pod \"glance-db-create-67phv\" (UID: \"29fb6825-d8d1-4401-b0ed-e1f3538cf4c5\") " pod="openstack/glance-db-create-67phv" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.520413 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5f4597a-1f01-4082-aada-cc7b976fad9a-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"e5f4597a-1f01-4082-aada-cc7b976fad9a\") " pod="openstack/ovn-northd-0" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.520434 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5f4597a-1f01-4082-aada-cc7b976fad9a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"e5f4597a-1f01-4082-aada-cc7b976fad9a\") " pod="openstack/ovn-northd-0" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.520460 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5f4597a-1f01-4082-aada-cc7b976fad9a-config\") pod \"ovn-northd-0\" (UID: \"e5f4597a-1f01-4082-aada-cc7b976fad9a\") " pod="openstack/ovn-northd-0" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.520479 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26ae359c-09e5-48c8-8f17-a5e6d7b75529-ovsdbserver-nb\") pod \"dnsmasq-dns-5d86d68bf7-mn47f\" (UID: \"26ae359c-09e5-48c8-8f17-a5e6d7b75529\") " pod="openstack/dnsmasq-dns-5d86d68bf7-mn47f" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.520499 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87xzb\" (UniqueName: \"kubernetes.io/projected/26ae359c-09e5-48c8-8f17-a5e6d7b75529-kube-api-access-87xzb\") pod \"dnsmasq-dns-5d86d68bf7-mn47f\" (UID: \"26ae359c-09e5-48c8-8f17-a5e6d7b75529\") " pod="openstack/dnsmasq-dns-5d86d68bf7-mn47f" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.520521 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5f4597a-1f01-4082-aada-cc7b976fad9a-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"e5f4597a-1f01-4082-aada-cc7b976fad9a\") " pod="openstack/ovn-northd-0" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.520540 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e5f4597a-1f01-4082-aada-cc7b976fad9a-scripts\") pod \"ovn-northd-0\" (UID: \"e5f4597a-1f01-4082-aada-cc7b976fad9a\") " pod="openstack/ovn-northd-0" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.520582 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wch2\" (UniqueName: \"kubernetes.io/projected/e5f4597a-1f01-4082-aada-cc7b976fad9a-kube-api-access-9wch2\") pod \"ovn-northd-0\" (UID: \"e5f4597a-1f01-4082-aada-cc7b976fad9a\") " pod="openstack/ovn-northd-0" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.520623 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26ae359c-09e5-48c8-8f17-a5e6d7b75529-ovsdbserver-sb\") pod \"dnsmasq-dns-5d86d68bf7-mn47f\" (UID: \"26ae359c-09e5-48c8-8f17-a5e6d7b75529\") " pod="openstack/dnsmasq-dns-5d86d68bf7-mn47f" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.520645 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e5f4597a-1f01-4082-aada-cc7b976fad9a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"e5f4597a-1f01-4082-aada-cc7b976fad9a\") " pod="openstack/ovn-northd-0" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.520663 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26ae359c-09e5-48c8-8f17-a5e6d7b75529-dns-svc\") pod \"dnsmasq-dns-5d86d68bf7-mn47f\" (UID: \"26ae359c-09e5-48c8-8f17-a5e6d7b75529\") " pod="openstack/dnsmasq-dns-5d86d68bf7-mn47f" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.522848 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-67phv"] Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.523484 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26ae359c-09e5-48c8-8f17-a5e6d7b75529-dns-svc\") pod \"dnsmasq-dns-5d86d68bf7-mn47f\" (UID: \"26ae359c-09e5-48c8-8f17-a5e6d7b75529\") " pod="openstack/dnsmasq-dns-5d86d68bf7-mn47f" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.523811 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26ae359c-09e5-48c8-8f17-a5e6d7b75529-ovsdbserver-nb\") pod \"dnsmasq-dns-5d86d68bf7-mn47f\" (UID: \"26ae359c-09e5-48c8-8f17-a5e6d7b75529\") " pod="openstack/dnsmasq-dns-5d86d68bf7-mn47f" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.523886 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26ae359c-09e5-48c8-8f17-a5e6d7b75529-config\") pod \"dnsmasq-dns-5d86d68bf7-mn47f\" (UID: \"26ae359c-09e5-48c8-8f17-a5e6d7b75529\") " pod="openstack/dnsmasq-dns-5d86d68bf7-mn47f" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.524300 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26ae359c-09e5-48c8-8f17-a5e6d7b75529-ovsdbserver-sb\") pod \"dnsmasq-dns-5d86d68bf7-mn47f\" (UID: \"26ae359c-09e5-48c8-8f17-a5e6d7b75529\") " pod="openstack/dnsmasq-dns-5d86d68bf7-mn47f" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.543471 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87xzb\" (UniqueName: \"kubernetes.io/projected/26ae359c-09e5-48c8-8f17-a5e6d7b75529-kube-api-access-87xzb\") pod \"dnsmasq-dns-5d86d68bf7-mn47f\" (UID: \"26ae359c-09e5-48c8-8f17-a5e6d7b75529\") " pod="openstack/dnsmasq-dns-5d86d68bf7-mn47f" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.585281 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-4lszz" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.625912 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5f4597a-1f01-4082-aada-cc7b976fad9a-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"e5f4597a-1f01-4082-aada-cc7b976fad9a\") " pod="openstack/ovn-northd-0" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.625961 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wgtg\" (UniqueName: \"kubernetes.io/projected/29fb6825-d8d1-4401-b0ed-e1f3538cf4c5-kube-api-access-9wgtg\") pod \"glance-db-create-67phv\" (UID: \"29fb6825-d8d1-4401-b0ed-e1f3538cf4c5\") " pod="openstack/glance-db-create-67phv" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.625986 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5f4597a-1f01-4082-aada-cc7b976fad9a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"e5f4597a-1f01-4082-aada-cc7b976fad9a\") " pod="openstack/ovn-northd-0" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.626019 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5f4597a-1f01-4082-aada-cc7b976fad9a-config\") pod \"ovn-northd-0\" (UID: \"e5f4597a-1f01-4082-aada-cc7b976fad9a\") " pod="openstack/ovn-northd-0" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.626042 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5f4597a-1f01-4082-aada-cc7b976fad9a-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"e5f4597a-1f01-4082-aada-cc7b976fad9a\") " pod="openstack/ovn-northd-0" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.626056 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e5f4597a-1f01-4082-aada-cc7b976fad9a-scripts\") pod \"ovn-northd-0\" (UID: \"e5f4597a-1f01-4082-aada-cc7b976fad9a\") " pod="openstack/ovn-northd-0" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.626072 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wch2\" (UniqueName: \"kubernetes.io/projected/e5f4597a-1f01-4082-aada-cc7b976fad9a-kube-api-access-9wch2\") pod \"ovn-northd-0\" (UID: \"e5f4597a-1f01-4082-aada-cc7b976fad9a\") " pod="openstack/ovn-northd-0" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.626195 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e5f4597a-1f01-4082-aada-cc7b976fad9a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"e5f4597a-1f01-4082-aada-cc7b976fad9a\") " pod="openstack/ovn-northd-0" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.628389 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e5f4597a-1f01-4082-aada-cc7b976fad9a-scripts\") pod \"ovn-northd-0\" (UID: \"e5f4597a-1f01-4082-aada-cc7b976fad9a\") " pod="openstack/ovn-northd-0" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.628930 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e5f4597a-1f01-4082-aada-cc7b976fad9a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"e5f4597a-1f01-4082-aada-cc7b976fad9a\") " pod="openstack/ovn-northd-0" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.629026 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5f4597a-1f01-4082-aada-cc7b976fad9a-config\") pod \"ovn-northd-0\" (UID: \"e5f4597a-1f01-4082-aada-cc7b976fad9a\") " pod="openstack/ovn-northd-0" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.679698 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5f4597a-1f01-4082-aada-cc7b976fad9a-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"e5f4597a-1f01-4082-aada-cc7b976fad9a\") " pod="openstack/ovn-northd-0" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.680046 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5f4597a-1f01-4082-aada-cc7b976fad9a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"e5f4597a-1f01-4082-aada-cc7b976fad9a\") " pod="openstack/ovn-northd-0" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.680701 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5f4597a-1f01-4082-aada-cc7b976fad9a-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"e5f4597a-1f01-4082-aada-cc7b976fad9a\") " pod="openstack/ovn-northd-0" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.696672 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wgtg\" (UniqueName: \"kubernetes.io/projected/29fb6825-d8d1-4401-b0ed-e1f3538cf4c5-kube-api-access-9wgtg\") pod \"glance-db-create-67phv\" (UID: \"29fb6825-d8d1-4401-b0ed-e1f3538cf4c5\") " pod="openstack/glance-db-create-67phv" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.708769 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d86d68bf7-mn47f" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.732443 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wch2\" (UniqueName: \"kubernetes.io/projected/e5f4597a-1f01-4082-aada-cc7b976fad9a-kube-api-access-9wch2\") pod \"ovn-northd-0\" (UID: \"e5f4597a-1f01-4082-aada-cc7b976fad9a\") " pod="openstack/ovn-northd-0" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.845436 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.874247 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86f865fc45-784xp" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.924516 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86f865fc45-784xp" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.964860 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 01 09:51:21 crc kubenswrapper[4787]: I1001 09:51:21.987463 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-67phv" Oct 01 09:51:22 crc kubenswrapper[4787]: I1001 09:51:22.038334 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23cec814-a12e-4b66-922a-ea061ed7e10a-config\") pod \"23cec814-a12e-4b66-922a-ea061ed7e10a\" (UID: \"23cec814-a12e-4b66-922a-ea061ed7e10a\") " Oct 01 09:51:22 crc kubenswrapper[4787]: I1001 09:51:22.038387 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/23cec814-a12e-4b66-922a-ea061ed7e10a-dns-svc\") pod \"23cec814-a12e-4b66-922a-ea061ed7e10a\" (UID: \"23cec814-a12e-4b66-922a-ea061ed7e10a\") " Oct 01 09:51:22 crc kubenswrapper[4787]: I1001 09:51:22.038422 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2t2f\" (UniqueName: \"kubernetes.io/projected/23cec814-a12e-4b66-922a-ea061ed7e10a-kube-api-access-s2t2f\") pod \"23cec814-a12e-4b66-922a-ea061ed7e10a\" (UID: \"23cec814-a12e-4b66-922a-ea061ed7e10a\") " Oct 01 09:51:22 crc kubenswrapper[4787]: I1001 09:51:22.038486 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/23cec814-a12e-4b66-922a-ea061ed7e10a-ovsdbserver-sb\") pod \"23cec814-a12e-4b66-922a-ea061ed7e10a\" (UID: \"23cec814-a12e-4b66-922a-ea061ed7e10a\") " Oct 01 09:51:22 crc kubenswrapper[4787]: I1001 09:51:22.039152 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23cec814-a12e-4b66-922a-ea061ed7e10a-config" (OuterVolumeSpecName: "config") pod "23cec814-a12e-4b66-922a-ea061ed7e10a" (UID: "23cec814-a12e-4b66-922a-ea061ed7e10a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:51:22 crc kubenswrapper[4787]: I1001 09:51:22.039410 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23cec814-a12e-4b66-922a-ea061ed7e10a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "23cec814-a12e-4b66-922a-ea061ed7e10a" (UID: "23cec814-a12e-4b66-922a-ea061ed7e10a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:51:22 crc kubenswrapper[4787]: I1001 09:51:22.040824 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23cec814-a12e-4b66-922a-ea061ed7e10a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "23cec814-a12e-4b66-922a-ea061ed7e10a" (UID: "23cec814-a12e-4b66-922a-ea061ed7e10a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:51:22 crc kubenswrapper[4787]: I1001 09:51:22.043738 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23cec814-a12e-4b66-922a-ea061ed7e10a-kube-api-access-s2t2f" (OuterVolumeSpecName: "kube-api-access-s2t2f") pod "23cec814-a12e-4b66-922a-ea061ed7e10a" (UID: "23cec814-a12e-4b66-922a-ea061ed7e10a"). InnerVolumeSpecName "kube-api-access-s2t2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:51:22 crc kubenswrapper[4787]: I1001 09:51:22.060121 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-d9cxp"] Oct 01 09:51:22 crc kubenswrapper[4787]: W1001 09:51:22.065837 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod11dbf4b0_9694_40e2_a122_9da928c50ad8.slice/crio-77fa0bcc557477734cd75ac8529df62838c334a30e39f166b481d46f382bb33e WatchSource:0}: Error finding container 77fa0bcc557477734cd75ac8529df62838c334a30e39f166b481d46f382bb33e: Status 404 returned error can't find the container with id 77fa0bcc557477734cd75ac8529df62838c334a30e39f166b481d46f382bb33e Oct 01 09:51:22 crc kubenswrapper[4787]: I1001 09:51:22.140542 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23cec814-a12e-4b66-922a-ea061ed7e10a-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:22 crc kubenswrapper[4787]: I1001 09:51:22.140570 4787 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/23cec814-a12e-4b66-922a-ea061ed7e10a-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:22 crc kubenswrapper[4787]: I1001 09:51:22.140580 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2t2f\" (UniqueName: \"kubernetes.io/projected/23cec814-a12e-4b66-922a-ea061ed7e10a-kube-api-access-s2t2f\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:22 crc kubenswrapper[4787]: I1001 09:51:22.140593 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/23cec814-a12e-4b66-922a-ea061ed7e10a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:22 crc kubenswrapper[4787]: I1001 09:51:22.279621 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-4lszz"] Oct 01 09:51:22 crc kubenswrapper[4787]: I1001 09:51:22.285626 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-67phv"] Oct 01 09:51:22 crc kubenswrapper[4787]: W1001 09:51:22.293216 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod11413a2c_48bf_4160_8e54_d843305b3b97.slice/crio-3446a6715489e8a95a92cd0a501948d8af8a0644c468b8c1483ac23f5de1e906 WatchSource:0}: Error finding container 3446a6715489e8a95a92cd0a501948d8af8a0644c468b8c1483ac23f5de1e906: Status 404 returned error can't find the container with id 3446a6715489e8a95a92cd0a501948d8af8a0644c468b8c1483ac23f5de1e906 Oct 01 09:51:22 crc kubenswrapper[4787]: I1001 09:51:22.338353 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d86d68bf7-mn47f"] Oct 01 09:51:22 crc kubenswrapper[4787]: W1001 09:51:22.343099 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod26ae359c_09e5_48c8_8f17_a5e6d7b75529.slice/crio-6e28307735a8067fdb0826521527ea73afa523cdd07f1d588e00e1789850ab5a WatchSource:0}: Error finding container 6e28307735a8067fdb0826521527ea73afa523cdd07f1d588e00e1789850ab5a: Status 404 returned error can't find the container with id 6e28307735a8067fdb0826521527ea73afa523cdd07f1d588e00e1789850ab5a Oct 01 09:51:22 crc kubenswrapper[4787]: I1001 09:51:22.396185 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 01 09:51:22 crc kubenswrapper[4787]: I1001 09:51:22.876865 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"e5f4597a-1f01-4082-aada-cc7b976fad9a","Type":"ContainerStarted","Data":"2824232a3495dd126a8c038836bad3d15a5ddc2f5aca08b13958f353ef4ac376"} Oct 01 09:51:22 crc kubenswrapper[4787]: I1001 09:51:22.879196 4787 generic.go:334] "Generic (PLEG): container finished" podID="26ae359c-09e5-48c8-8f17-a5e6d7b75529" containerID="0bbd6dfd786381be1dc89832debdc18140f5a5d66be20be5f53c17822216812a" exitCode=0 Oct 01 09:51:22 crc kubenswrapper[4787]: I1001 09:51:22.879251 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d86d68bf7-mn47f" event={"ID":"26ae359c-09e5-48c8-8f17-a5e6d7b75529","Type":"ContainerDied","Data":"0bbd6dfd786381be1dc89832debdc18140f5a5d66be20be5f53c17822216812a"} Oct 01 09:51:22 crc kubenswrapper[4787]: I1001 09:51:22.879269 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d86d68bf7-mn47f" event={"ID":"26ae359c-09e5-48c8-8f17-a5e6d7b75529","Type":"ContainerStarted","Data":"6e28307735a8067fdb0826521527ea73afa523cdd07f1d588e00e1789850ab5a"} Oct 01 09:51:22 crc kubenswrapper[4787]: I1001 09:51:22.881463 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-d9cxp" event={"ID":"11dbf4b0-9694-40e2-a122-9da928c50ad8","Type":"ContainerStarted","Data":"31dbe1ab90889b6ecac2a53e68bdd5c82bad6c2fd5316a644bede96fe3a27f87"} Oct 01 09:51:22 crc kubenswrapper[4787]: I1001 09:51:22.881492 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-d9cxp" event={"ID":"11dbf4b0-9694-40e2-a122-9da928c50ad8","Type":"ContainerStarted","Data":"77fa0bcc557477734cd75ac8529df62838c334a30e39f166b481d46f382bb33e"} Oct 01 09:51:22 crc kubenswrapper[4787]: I1001 09:51:22.883869 4787 generic.go:334] "Generic (PLEG): container finished" podID="29fb6825-d8d1-4401-b0ed-e1f3538cf4c5" containerID="3f08c2d08d5b370333d23fbf069dd296eb3b2296c94e98ef187f5b7c07f3ed23" exitCode=0 Oct 01 09:51:22 crc kubenswrapper[4787]: I1001 09:51:22.883911 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-67phv" event={"ID":"29fb6825-d8d1-4401-b0ed-e1f3538cf4c5","Type":"ContainerDied","Data":"3f08c2d08d5b370333d23fbf069dd296eb3b2296c94e98ef187f5b7c07f3ed23"} Oct 01 09:51:22 crc kubenswrapper[4787]: I1001 09:51:22.883926 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-67phv" event={"ID":"29fb6825-d8d1-4401-b0ed-e1f3538cf4c5","Type":"ContainerStarted","Data":"9483e2e2d026d57b1521d10d57383c389bfe24c440c9086982ef3f2682d18d3c"} Oct 01 09:51:22 crc kubenswrapper[4787]: I1001 09:51:22.888679 4787 generic.go:334] "Generic (PLEG): container finished" podID="11413a2c-48bf-4160-8e54-d843305b3b97" containerID="a683ee182b63428f19a3d11e7554ca1a4e3f554728ce4dd7334e736c8d8dd617" exitCode=0 Oct 01 09:51:22 crc kubenswrapper[4787]: I1001 09:51:22.888781 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86f865fc45-784xp" Oct 01 09:51:22 crc kubenswrapper[4787]: I1001 09:51:22.889238 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-4lszz" event={"ID":"11413a2c-48bf-4160-8e54-d843305b3b97","Type":"ContainerDied","Data":"a683ee182b63428f19a3d11e7554ca1a4e3f554728ce4dd7334e736c8d8dd617"} Oct 01 09:51:22 crc kubenswrapper[4787]: I1001 09:51:22.889293 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-4lszz" event={"ID":"11413a2c-48bf-4160-8e54-d843305b3b97","Type":"ContainerStarted","Data":"3446a6715489e8a95a92cd0a501948d8af8a0644c468b8c1483ac23f5de1e906"} Oct 01 09:51:22 crc kubenswrapper[4787]: I1001 09:51:22.976629 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-d9cxp" podStartSLOduration=1.976613312 podStartE2EDuration="1.976613312s" podCreationTimestamp="2025-10-01 09:51:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:51:22.955771938 +0000 UTC m=+915.070916105" watchObservedRunningTime="2025-10-01 09:51:22.976613312 +0000 UTC m=+915.091757469" Oct 01 09:51:22 crc kubenswrapper[4787]: I1001 09:51:22.994934 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86f865fc45-784xp"] Oct 01 09:51:23 crc kubenswrapper[4787]: I1001 09:51:23.003917 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86f865fc45-784xp"] Oct 01 09:51:23 crc kubenswrapper[4787]: I1001 09:51:23.320947 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d86d68bf7-mn47f"] Oct 01 09:51:23 crc kubenswrapper[4787]: I1001 09:51:23.388317 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6c6d5d5bd7-4mbvj"] Oct 01 09:51:23 crc kubenswrapper[4787]: I1001 09:51:23.411335 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c6d5d5bd7-4mbvj" Oct 01 09:51:23 crc kubenswrapper[4787]: I1001 09:51:23.442634 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c6d5d5bd7-4mbvj"] Oct 01 09:51:23 crc kubenswrapper[4787]: I1001 09:51:23.468035 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bbade729-c3cb-447b-bdd7-251fbc80c454-dns-svc\") pod \"dnsmasq-dns-6c6d5d5bd7-4mbvj\" (UID: \"bbade729-c3cb-447b-bdd7-251fbc80c454\") " pod="openstack/dnsmasq-dns-6c6d5d5bd7-4mbvj" Oct 01 09:51:23 crc kubenswrapper[4787]: I1001 09:51:23.468124 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdp72\" (UniqueName: \"kubernetes.io/projected/bbade729-c3cb-447b-bdd7-251fbc80c454-kube-api-access-wdp72\") pod \"dnsmasq-dns-6c6d5d5bd7-4mbvj\" (UID: \"bbade729-c3cb-447b-bdd7-251fbc80c454\") " pod="openstack/dnsmasq-dns-6c6d5d5bd7-4mbvj" Oct 01 09:51:23 crc kubenswrapper[4787]: I1001 09:51:23.468183 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bbade729-c3cb-447b-bdd7-251fbc80c454-config\") pod \"dnsmasq-dns-6c6d5d5bd7-4mbvj\" (UID: \"bbade729-c3cb-447b-bdd7-251fbc80c454\") " pod="openstack/dnsmasq-dns-6c6d5d5bd7-4mbvj" Oct 01 09:51:23 crc kubenswrapper[4787]: I1001 09:51:23.468208 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bbade729-c3cb-447b-bdd7-251fbc80c454-ovsdbserver-nb\") pod \"dnsmasq-dns-6c6d5d5bd7-4mbvj\" (UID: \"bbade729-c3cb-447b-bdd7-251fbc80c454\") " pod="openstack/dnsmasq-dns-6c6d5d5bd7-4mbvj" Oct 01 09:51:23 crc kubenswrapper[4787]: I1001 09:51:23.468234 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bbade729-c3cb-447b-bdd7-251fbc80c454-ovsdbserver-sb\") pod \"dnsmasq-dns-6c6d5d5bd7-4mbvj\" (UID: \"bbade729-c3cb-447b-bdd7-251fbc80c454\") " pod="openstack/dnsmasq-dns-6c6d5d5bd7-4mbvj" Oct 01 09:51:23 crc kubenswrapper[4787]: I1001 09:51:23.571199 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bbade729-c3cb-447b-bdd7-251fbc80c454-dns-svc\") pod \"dnsmasq-dns-6c6d5d5bd7-4mbvj\" (UID: \"bbade729-c3cb-447b-bdd7-251fbc80c454\") " pod="openstack/dnsmasq-dns-6c6d5d5bd7-4mbvj" Oct 01 09:51:23 crc kubenswrapper[4787]: I1001 09:51:23.571269 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdp72\" (UniqueName: \"kubernetes.io/projected/bbade729-c3cb-447b-bdd7-251fbc80c454-kube-api-access-wdp72\") pod \"dnsmasq-dns-6c6d5d5bd7-4mbvj\" (UID: \"bbade729-c3cb-447b-bdd7-251fbc80c454\") " pod="openstack/dnsmasq-dns-6c6d5d5bd7-4mbvj" Oct 01 09:51:23 crc kubenswrapper[4787]: I1001 09:51:23.571332 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bbade729-c3cb-447b-bdd7-251fbc80c454-config\") pod \"dnsmasq-dns-6c6d5d5bd7-4mbvj\" (UID: \"bbade729-c3cb-447b-bdd7-251fbc80c454\") " pod="openstack/dnsmasq-dns-6c6d5d5bd7-4mbvj" Oct 01 09:51:23 crc kubenswrapper[4787]: I1001 09:51:23.571357 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bbade729-c3cb-447b-bdd7-251fbc80c454-ovsdbserver-nb\") pod \"dnsmasq-dns-6c6d5d5bd7-4mbvj\" (UID: \"bbade729-c3cb-447b-bdd7-251fbc80c454\") " pod="openstack/dnsmasq-dns-6c6d5d5bd7-4mbvj" Oct 01 09:51:23 crc kubenswrapper[4787]: I1001 09:51:23.571385 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bbade729-c3cb-447b-bdd7-251fbc80c454-ovsdbserver-sb\") pod \"dnsmasq-dns-6c6d5d5bd7-4mbvj\" (UID: \"bbade729-c3cb-447b-bdd7-251fbc80c454\") " pod="openstack/dnsmasq-dns-6c6d5d5bd7-4mbvj" Oct 01 09:51:23 crc kubenswrapper[4787]: I1001 09:51:23.572235 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bbade729-c3cb-447b-bdd7-251fbc80c454-dns-svc\") pod \"dnsmasq-dns-6c6d5d5bd7-4mbvj\" (UID: \"bbade729-c3cb-447b-bdd7-251fbc80c454\") " pod="openstack/dnsmasq-dns-6c6d5d5bd7-4mbvj" Oct 01 09:51:23 crc kubenswrapper[4787]: I1001 09:51:23.572528 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bbade729-c3cb-447b-bdd7-251fbc80c454-config\") pod \"dnsmasq-dns-6c6d5d5bd7-4mbvj\" (UID: \"bbade729-c3cb-447b-bdd7-251fbc80c454\") " pod="openstack/dnsmasq-dns-6c6d5d5bd7-4mbvj" Oct 01 09:51:23 crc kubenswrapper[4787]: I1001 09:51:23.572740 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bbade729-c3cb-447b-bdd7-251fbc80c454-ovsdbserver-sb\") pod \"dnsmasq-dns-6c6d5d5bd7-4mbvj\" (UID: \"bbade729-c3cb-447b-bdd7-251fbc80c454\") " pod="openstack/dnsmasq-dns-6c6d5d5bd7-4mbvj" Oct 01 09:51:23 crc kubenswrapper[4787]: I1001 09:51:23.572810 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bbade729-c3cb-447b-bdd7-251fbc80c454-ovsdbserver-nb\") pod \"dnsmasq-dns-6c6d5d5bd7-4mbvj\" (UID: \"bbade729-c3cb-447b-bdd7-251fbc80c454\") " pod="openstack/dnsmasq-dns-6c6d5d5bd7-4mbvj" Oct 01 09:51:23 crc kubenswrapper[4787]: I1001 09:51:23.605205 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdp72\" (UniqueName: \"kubernetes.io/projected/bbade729-c3cb-447b-bdd7-251fbc80c454-kube-api-access-wdp72\") pod \"dnsmasq-dns-6c6d5d5bd7-4mbvj\" (UID: \"bbade729-c3cb-447b-bdd7-251fbc80c454\") " pod="openstack/dnsmasq-dns-6c6d5d5bd7-4mbvj" Oct 01 09:51:23 crc kubenswrapper[4787]: I1001 09:51:23.754486 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c6d5d5bd7-4mbvj" Oct 01 09:51:23 crc kubenswrapper[4787]: I1001 09:51:23.896815 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d86d68bf7-mn47f" event={"ID":"26ae359c-09e5-48c8-8f17-a5e6d7b75529","Type":"ContainerStarted","Data":"9312726f8da9cacc2afb441fb3742aa3356b2987cc2f23c6192836a2fa7d2e18"} Oct 01 09:51:23 crc kubenswrapper[4787]: I1001 09:51:23.922852 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5d86d68bf7-mn47f" podStartSLOduration=2.922834951 podStartE2EDuration="2.922834951s" podCreationTimestamp="2025-10-01 09:51:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:51:23.91929601 +0000 UTC m=+916.034440157" watchObservedRunningTime="2025-10-01 09:51:23.922834951 +0000 UTC m=+916.037979108" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.301708 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-67phv" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.304665 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-4lszz" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.383451 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nd7pk\" (UniqueName: \"kubernetes.io/projected/11413a2c-48bf-4160-8e54-d843305b3b97-kube-api-access-nd7pk\") pod \"11413a2c-48bf-4160-8e54-d843305b3b97\" (UID: \"11413a2c-48bf-4160-8e54-d843305b3b97\") " Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.384791 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wgtg\" (UniqueName: \"kubernetes.io/projected/29fb6825-d8d1-4401-b0ed-e1f3538cf4c5-kube-api-access-9wgtg\") pod \"29fb6825-d8d1-4401-b0ed-e1f3538cf4c5\" (UID: \"29fb6825-d8d1-4401-b0ed-e1f3538cf4c5\") " Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.389099 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11413a2c-48bf-4160-8e54-d843305b3b97-kube-api-access-nd7pk" (OuterVolumeSpecName: "kube-api-access-nd7pk") pod "11413a2c-48bf-4160-8e54-d843305b3b97" (UID: "11413a2c-48bf-4160-8e54-d843305b3b97"). InnerVolumeSpecName "kube-api-access-nd7pk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.393148 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29fb6825-d8d1-4401-b0ed-e1f3538cf4c5-kube-api-access-9wgtg" (OuterVolumeSpecName: "kube-api-access-9wgtg") pod "29fb6825-d8d1-4401-b0ed-e1f3538cf4c5" (UID: "29fb6825-d8d1-4401-b0ed-e1f3538cf4c5"). InnerVolumeSpecName "kube-api-access-9wgtg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.418668 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 01 09:51:24 crc kubenswrapper[4787]: E1001 09:51:24.419048 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11413a2c-48bf-4160-8e54-d843305b3b97" containerName="mariadb-database-create" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.419093 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="11413a2c-48bf-4160-8e54-d843305b3b97" containerName="mariadb-database-create" Oct 01 09:51:24 crc kubenswrapper[4787]: E1001 09:51:24.419123 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29fb6825-d8d1-4401-b0ed-e1f3538cf4c5" containerName="mariadb-database-create" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.419129 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="29fb6825-d8d1-4401-b0ed-e1f3538cf4c5" containerName="mariadb-database-create" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.419288 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="29fb6825-d8d1-4401-b0ed-e1f3538cf4c5" containerName="mariadb-database-create" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.419312 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="11413a2c-48bf-4160-8e54-d843305b3b97" containerName="mariadb-database-create" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.424324 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.427300 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.427357 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.427805 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-zsgkj" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.428131 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.440712 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.486554 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/3432bc16-99b4-431a-aeb1-600e826bbc3e-cache\") pod \"swift-storage-0\" (UID: \"3432bc16-99b4-431a-aeb1-600e826bbc3e\") " pod="openstack/swift-storage-0" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.486607 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/3432bc16-99b4-431a-aeb1-600e826bbc3e-lock\") pod \"swift-storage-0\" (UID: \"3432bc16-99b4-431a-aeb1-600e826bbc3e\") " pod="openstack/swift-storage-0" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.486666 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzdqq\" (UniqueName: \"kubernetes.io/projected/3432bc16-99b4-431a-aeb1-600e826bbc3e-kube-api-access-vzdqq\") pod \"swift-storage-0\" (UID: \"3432bc16-99b4-431a-aeb1-600e826bbc3e\") " pod="openstack/swift-storage-0" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.486692 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3432bc16-99b4-431a-aeb1-600e826bbc3e-etc-swift\") pod \"swift-storage-0\" (UID: \"3432bc16-99b4-431a-aeb1-600e826bbc3e\") " pod="openstack/swift-storage-0" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.486711 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"3432bc16-99b4-431a-aeb1-600e826bbc3e\") " pod="openstack/swift-storage-0" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.486797 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nd7pk\" (UniqueName: \"kubernetes.io/projected/11413a2c-48bf-4160-8e54-d843305b3b97-kube-api-access-nd7pk\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.486810 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wgtg\" (UniqueName: \"kubernetes.io/projected/29fb6825-d8d1-4401-b0ed-e1f3538cf4c5-kube-api-access-9wgtg\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.538669 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23cec814-a12e-4b66-922a-ea061ed7e10a" path="/var/lib/kubelet/pods/23cec814-a12e-4b66-922a-ea061ed7e10a/volumes" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.587937 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/3432bc16-99b4-431a-aeb1-600e826bbc3e-cache\") pod \"swift-storage-0\" (UID: \"3432bc16-99b4-431a-aeb1-600e826bbc3e\") " pod="openstack/swift-storage-0" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.590006 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/3432bc16-99b4-431a-aeb1-600e826bbc3e-lock\") pod \"swift-storage-0\" (UID: \"3432bc16-99b4-431a-aeb1-600e826bbc3e\") " pod="openstack/swift-storage-0" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.590682 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzdqq\" (UniqueName: \"kubernetes.io/projected/3432bc16-99b4-431a-aeb1-600e826bbc3e-kube-api-access-vzdqq\") pod \"swift-storage-0\" (UID: \"3432bc16-99b4-431a-aeb1-600e826bbc3e\") " pod="openstack/swift-storage-0" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.591519 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3432bc16-99b4-431a-aeb1-600e826bbc3e-etc-swift\") pod \"swift-storage-0\" (UID: \"3432bc16-99b4-431a-aeb1-600e826bbc3e\") " pod="openstack/swift-storage-0" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.591691 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"3432bc16-99b4-431a-aeb1-600e826bbc3e\") " pod="openstack/swift-storage-0" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.590535 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/3432bc16-99b4-431a-aeb1-600e826bbc3e-lock\") pod \"swift-storage-0\" (UID: \"3432bc16-99b4-431a-aeb1-600e826bbc3e\") " pod="openstack/swift-storage-0" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.589783 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/3432bc16-99b4-431a-aeb1-600e826bbc3e-cache\") pod \"swift-storage-0\" (UID: \"3432bc16-99b4-431a-aeb1-600e826bbc3e\") " pod="openstack/swift-storage-0" Oct 01 09:51:24 crc kubenswrapper[4787]: E1001 09:51:24.591781 4787 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 01 09:51:24 crc kubenswrapper[4787]: E1001 09:51:24.592285 4787 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.592020 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"3432bc16-99b4-431a-aeb1-600e826bbc3e\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/swift-storage-0" Oct 01 09:51:24 crc kubenswrapper[4787]: E1001 09:51:24.592509 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3432bc16-99b4-431a-aeb1-600e826bbc3e-etc-swift podName:3432bc16-99b4-431a-aeb1-600e826bbc3e nodeName:}" failed. No retries permitted until 2025-10-01 09:51:25.092495172 +0000 UTC m=+917.207639329 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/3432bc16-99b4-431a-aeb1-600e826bbc3e-etc-swift") pod "swift-storage-0" (UID: "3432bc16-99b4-431a-aeb1-600e826bbc3e") : configmap "swift-ring-files" not found Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.613731 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzdqq\" (UniqueName: \"kubernetes.io/projected/3432bc16-99b4-431a-aeb1-600e826bbc3e-kube-api-access-vzdqq\") pod \"swift-storage-0\" (UID: \"3432bc16-99b4-431a-aeb1-600e826bbc3e\") " pod="openstack/swift-storage-0" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.617066 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"3432bc16-99b4-431a-aeb1-600e826bbc3e\") " pod="openstack/swift-storage-0" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.638379 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c6d5d5bd7-4mbvj"] Oct 01 09:51:24 crc kubenswrapper[4787]: W1001 09:51:24.641260 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbbade729_c3cb_447b_bdd7_251fbc80c454.slice/crio-686aeae305b7a63e69cd2cdf66eb516ba5c4667697ac7753b580e525d09e7cea WatchSource:0}: Error finding container 686aeae305b7a63e69cd2cdf66eb516ba5c4667697ac7753b580e525d09e7cea: Status 404 returned error can't find the container with id 686aeae305b7a63e69cd2cdf66eb516ba5c4667697ac7753b580e525d09e7cea Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.916266 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-67phv" event={"ID":"29fb6825-d8d1-4401-b0ed-e1f3538cf4c5","Type":"ContainerDied","Data":"9483e2e2d026d57b1521d10d57383c389bfe24c440c9086982ef3f2682d18d3c"} Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.917287 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9483e2e2d026d57b1521d10d57383c389bfe24c440c9086982ef3f2682d18d3c" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.916521 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-67phv" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.921191 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-4lszz" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.921199 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-4lszz" event={"ID":"11413a2c-48bf-4160-8e54-d843305b3b97","Type":"ContainerDied","Data":"3446a6715489e8a95a92cd0a501948d8af8a0644c468b8c1483ac23f5de1e906"} Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.921297 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3446a6715489e8a95a92cd0a501948d8af8a0644c468b8c1483ac23f5de1e906" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.925480 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"e5f4597a-1f01-4082-aada-cc7b976fad9a","Type":"ContainerStarted","Data":"d47d66af3d540056d368a28e4a8ea0b8fc0a8c922fdaf88fa4f5b296d2dd4ade"} Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.925548 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"e5f4597a-1f01-4082-aada-cc7b976fad9a","Type":"ContainerStarted","Data":"5fbc02ac392dabca2d4251c62bcf7a7d99bfd140e929301a3b66f36ee24d9e0b"} Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.925610 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.928002 4787 generic.go:334] "Generic (PLEG): container finished" podID="bbade729-c3cb-447b-bdd7-251fbc80c454" containerID="6f2178ce4b22d3637acac7d52e9173e5f851ab77726fba536ab7ec09e3872032" exitCode=0 Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.928096 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c6d5d5bd7-4mbvj" event={"ID":"bbade729-c3cb-447b-bdd7-251fbc80c454","Type":"ContainerDied","Data":"6f2178ce4b22d3637acac7d52e9173e5f851ab77726fba536ab7ec09e3872032"} Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.928150 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c6d5d5bd7-4mbvj" event={"ID":"bbade729-c3cb-447b-bdd7-251fbc80c454","Type":"ContainerStarted","Data":"686aeae305b7a63e69cd2cdf66eb516ba5c4667697ac7753b580e525d09e7cea"} Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.928478 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5d86d68bf7-mn47f" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.928459 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5d86d68bf7-mn47f" podUID="26ae359c-09e5-48c8-8f17-a5e6d7b75529" containerName="dnsmasq-dns" containerID="cri-o://9312726f8da9cacc2afb441fb3742aa3356b2987cc2f23c6192836a2fa7d2e18" gracePeriod=10 Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.966599 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-kkdkg"] Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.968479 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-kkdkg" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.974580 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.977966 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.978517 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.994048 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-kkdkg"] Oct 01 09:51:24 crc kubenswrapper[4787]: I1001 09:51:24.996474 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.259336358 podStartE2EDuration="3.996451274s" podCreationTimestamp="2025-10-01 09:51:21 +0000 UTC" firstStartedPulling="2025-10-01 09:51:22.416050167 +0000 UTC m=+914.531194324" lastFinishedPulling="2025-10-01 09:51:24.153165083 +0000 UTC m=+916.268309240" observedRunningTime="2025-10-01 09:51:24.982542828 +0000 UTC m=+917.097686995" watchObservedRunningTime="2025-10-01 09:51:24.996451274 +0000 UTC m=+917.111595431" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.104708 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqg4v\" (UniqueName: \"kubernetes.io/projected/ed714165-98db-45a6-b96b-69601bff3e56-kube-api-access-sqg4v\") pod \"swift-ring-rebalance-kkdkg\" (UID: \"ed714165-98db-45a6-b96b-69601bff3e56\") " pod="openstack/swift-ring-rebalance-kkdkg" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.104771 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ed714165-98db-45a6-b96b-69601bff3e56-etc-swift\") pod \"swift-ring-rebalance-kkdkg\" (UID: \"ed714165-98db-45a6-b96b-69601bff3e56\") " pod="openstack/swift-ring-rebalance-kkdkg" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.104807 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ed714165-98db-45a6-b96b-69601bff3e56-swiftconf\") pod \"swift-ring-rebalance-kkdkg\" (UID: \"ed714165-98db-45a6-b96b-69601bff3e56\") " pod="openstack/swift-ring-rebalance-kkdkg" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.104944 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ed714165-98db-45a6-b96b-69601bff3e56-scripts\") pod \"swift-ring-rebalance-kkdkg\" (UID: \"ed714165-98db-45a6-b96b-69601bff3e56\") " pod="openstack/swift-ring-rebalance-kkdkg" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.104993 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ed714165-98db-45a6-b96b-69601bff3e56-dispersionconf\") pod \"swift-ring-rebalance-kkdkg\" (UID: \"ed714165-98db-45a6-b96b-69601bff3e56\") " pod="openstack/swift-ring-rebalance-kkdkg" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.105204 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3432bc16-99b4-431a-aeb1-600e826bbc3e-etc-swift\") pod \"swift-storage-0\" (UID: \"3432bc16-99b4-431a-aeb1-600e826bbc3e\") " pod="openstack/swift-storage-0" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.105304 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ed714165-98db-45a6-b96b-69601bff3e56-ring-data-devices\") pod \"swift-ring-rebalance-kkdkg\" (UID: \"ed714165-98db-45a6-b96b-69601bff3e56\") " pod="openstack/swift-ring-rebalance-kkdkg" Oct 01 09:51:25 crc kubenswrapper[4787]: E1001 09:51:25.105425 4787 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 01 09:51:25 crc kubenswrapper[4787]: E1001 09:51:25.105445 4787 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.105448 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed714165-98db-45a6-b96b-69601bff3e56-combined-ca-bundle\") pod \"swift-ring-rebalance-kkdkg\" (UID: \"ed714165-98db-45a6-b96b-69601bff3e56\") " pod="openstack/swift-ring-rebalance-kkdkg" Oct 01 09:51:25 crc kubenswrapper[4787]: E1001 09:51:25.105494 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3432bc16-99b4-431a-aeb1-600e826bbc3e-etc-swift podName:3432bc16-99b4-431a-aeb1-600e826bbc3e nodeName:}" failed. No retries permitted until 2025-10-01 09:51:26.105476747 +0000 UTC m=+918.220620904 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/3432bc16-99b4-431a-aeb1-600e826bbc3e-etc-swift") pod "swift-storage-0" (UID: "3432bc16-99b4-431a-aeb1-600e826bbc3e") : configmap "swift-ring-files" not found Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.207034 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ed714165-98db-45a6-b96b-69601bff3e56-ring-data-devices\") pod \"swift-ring-rebalance-kkdkg\" (UID: \"ed714165-98db-45a6-b96b-69601bff3e56\") " pod="openstack/swift-ring-rebalance-kkdkg" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.207125 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed714165-98db-45a6-b96b-69601bff3e56-combined-ca-bundle\") pod \"swift-ring-rebalance-kkdkg\" (UID: \"ed714165-98db-45a6-b96b-69601bff3e56\") " pod="openstack/swift-ring-rebalance-kkdkg" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.207146 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqg4v\" (UniqueName: \"kubernetes.io/projected/ed714165-98db-45a6-b96b-69601bff3e56-kube-api-access-sqg4v\") pod \"swift-ring-rebalance-kkdkg\" (UID: \"ed714165-98db-45a6-b96b-69601bff3e56\") " pod="openstack/swift-ring-rebalance-kkdkg" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.207177 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ed714165-98db-45a6-b96b-69601bff3e56-etc-swift\") pod \"swift-ring-rebalance-kkdkg\" (UID: \"ed714165-98db-45a6-b96b-69601bff3e56\") " pod="openstack/swift-ring-rebalance-kkdkg" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.207205 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ed714165-98db-45a6-b96b-69601bff3e56-swiftconf\") pod \"swift-ring-rebalance-kkdkg\" (UID: \"ed714165-98db-45a6-b96b-69601bff3e56\") " pod="openstack/swift-ring-rebalance-kkdkg" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.207242 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ed714165-98db-45a6-b96b-69601bff3e56-scripts\") pod \"swift-ring-rebalance-kkdkg\" (UID: \"ed714165-98db-45a6-b96b-69601bff3e56\") " pod="openstack/swift-ring-rebalance-kkdkg" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.207257 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ed714165-98db-45a6-b96b-69601bff3e56-dispersionconf\") pod \"swift-ring-rebalance-kkdkg\" (UID: \"ed714165-98db-45a6-b96b-69601bff3e56\") " pod="openstack/swift-ring-rebalance-kkdkg" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.209227 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ed714165-98db-45a6-b96b-69601bff3e56-ring-data-devices\") pod \"swift-ring-rebalance-kkdkg\" (UID: \"ed714165-98db-45a6-b96b-69601bff3e56\") " pod="openstack/swift-ring-rebalance-kkdkg" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.209701 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ed714165-98db-45a6-b96b-69601bff3e56-scripts\") pod \"swift-ring-rebalance-kkdkg\" (UID: \"ed714165-98db-45a6-b96b-69601bff3e56\") " pod="openstack/swift-ring-rebalance-kkdkg" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.209923 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ed714165-98db-45a6-b96b-69601bff3e56-etc-swift\") pod \"swift-ring-rebalance-kkdkg\" (UID: \"ed714165-98db-45a6-b96b-69601bff3e56\") " pod="openstack/swift-ring-rebalance-kkdkg" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.213605 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed714165-98db-45a6-b96b-69601bff3e56-combined-ca-bundle\") pod \"swift-ring-rebalance-kkdkg\" (UID: \"ed714165-98db-45a6-b96b-69601bff3e56\") " pod="openstack/swift-ring-rebalance-kkdkg" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.214642 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ed714165-98db-45a6-b96b-69601bff3e56-swiftconf\") pod \"swift-ring-rebalance-kkdkg\" (UID: \"ed714165-98db-45a6-b96b-69601bff3e56\") " pod="openstack/swift-ring-rebalance-kkdkg" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.220483 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ed714165-98db-45a6-b96b-69601bff3e56-dispersionconf\") pod \"swift-ring-rebalance-kkdkg\" (UID: \"ed714165-98db-45a6-b96b-69601bff3e56\") " pod="openstack/swift-ring-rebalance-kkdkg" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.228419 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqg4v\" (UniqueName: \"kubernetes.io/projected/ed714165-98db-45a6-b96b-69601bff3e56-kube-api-access-sqg4v\") pod \"swift-ring-rebalance-kkdkg\" (UID: \"ed714165-98db-45a6-b96b-69601bff3e56\") " pod="openstack/swift-ring-rebalance-kkdkg" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.315130 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d86d68bf7-mn47f" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.354649 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-kkdkg" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.410295 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87xzb\" (UniqueName: \"kubernetes.io/projected/26ae359c-09e5-48c8-8f17-a5e6d7b75529-kube-api-access-87xzb\") pod \"26ae359c-09e5-48c8-8f17-a5e6d7b75529\" (UID: \"26ae359c-09e5-48c8-8f17-a5e6d7b75529\") " Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.410429 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26ae359c-09e5-48c8-8f17-a5e6d7b75529-dns-svc\") pod \"26ae359c-09e5-48c8-8f17-a5e6d7b75529\" (UID: \"26ae359c-09e5-48c8-8f17-a5e6d7b75529\") " Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.410528 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26ae359c-09e5-48c8-8f17-a5e6d7b75529-ovsdbserver-nb\") pod \"26ae359c-09e5-48c8-8f17-a5e6d7b75529\" (UID: \"26ae359c-09e5-48c8-8f17-a5e6d7b75529\") " Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.410557 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26ae359c-09e5-48c8-8f17-a5e6d7b75529-config\") pod \"26ae359c-09e5-48c8-8f17-a5e6d7b75529\" (UID: \"26ae359c-09e5-48c8-8f17-a5e6d7b75529\") " Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.410649 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26ae359c-09e5-48c8-8f17-a5e6d7b75529-ovsdbserver-sb\") pod \"26ae359c-09e5-48c8-8f17-a5e6d7b75529\" (UID: \"26ae359c-09e5-48c8-8f17-a5e6d7b75529\") " Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.417572 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26ae359c-09e5-48c8-8f17-a5e6d7b75529-kube-api-access-87xzb" (OuterVolumeSpecName: "kube-api-access-87xzb") pod "26ae359c-09e5-48c8-8f17-a5e6d7b75529" (UID: "26ae359c-09e5-48c8-8f17-a5e6d7b75529"). InnerVolumeSpecName "kube-api-access-87xzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.463927 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26ae359c-09e5-48c8-8f17-a5e6d7b75529-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "26ae359c-09e5-48c8-8f17-a5e6d7b75529" (UID: "26ae359c-09e5-48c8-8f17-a5e6d7b75529"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.465384 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26ae359c-09e5-48c8-8f17-a5e6d7b75529-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "26ae359c-09e5-48c8-8f17-a5e6d7b75529" (UID: "26ae359c-09e5-48c8-8f17-a5e6d7b75529"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.479282 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26ae359c-09e5-48c8-8f17-a5e6d7b75529-config" (OuterVolumeSpecName: "config") pod "26ae359c-09e5-48c8-8f17-a5e6d7b75529" (UID: "26ae359c-09e5-48c8-8f17-a5e6d7b75529"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.490753 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26ae359c-09e5-48c8-8f17-a5e6d7b75529-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "26ae359c-09e5-48c8-8f17-a5e6d7b75529" (UID: "26ae359c-09e5-48c8-8f17-a5e6d7b75529"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.512910 4787 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26ae359c-09e5-48c8-8f17-a5e6d7b75529-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.512949 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26ae359c-09e5-48c8-8f17-a5e6d7b75529-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.512963 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26ae359c-09e5-48c8-8f17-a5e6d7b75529-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.512975 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26ae359c-09e5-48c8-8f17-a5e6d7b75529-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.512988 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87xzb\" (UniqueName: \"kubernetes.io/projected/26ae359c-09e5-48c8-8f17-a5e6d7b75529-kube-api-access-87xzb\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.853856 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-kkdkg"] Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.938355 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c6d5d5bd7-4mbvj" event={"ID":"bbade729-c3cb-447b-bdd7-251fbc80c454","Type":"ContainerStarted","Data":"6020b30fd47f95aa07a6b93dfa13052657b3120e459f6fe4674a14f5b6486a9c"} Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.938738 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6c6d5d5bd7-4mbvj" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.940898 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-kkdkg" event={"ID":"ed714165-98db-45a6-b96b-69601bff3e56","Type":"ContainerStarted","Data":"298e605666808eb9c6180bfe0fb573ee731369748b6db8ea79a1150ee8ec384e"} Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.943317 4787 generic.go:334] "Generic (PLEG): container finished" podID="26ae359c-09e5-48c8-8f17-a5e6d7b75529" containerID="9312726f8da9cacc2afb441fb3742aa3356b2987cc2f23c6192836a2fa7d2e18" exitCode=0 Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.943362 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d86d68bf7-mn47f" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.943423 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d86d68bf7-mn47f" event={"ID":"26ae359c-09e5-48c8-8f17-a5e6d7b75529","Type":"ContainerDied","Data":"9312726f8da9cacc2afb441fb3742aa3356b2987cc2f23c6192836a2fa7d2e18"} Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.943472 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d86d68bf7-mn47f" event={"ID":"26ae359c-09e5-48c8-8f17-a5e6d7b75529","Type":"ContainerDied","Data":"6e28307735a8067fdb0826521527ea73afa523cdd07f1d588e00e1789850ab5a"} Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.943517 4787 scope.go:117] "RemoveContainer" containerID="9312726f8da9cacc2afb441fb3742aa3356b2987cc2f23c6192836a2fa7d2e18" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.967556 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6c6d5d5bd7-4mbvj" podStartSLOduration=2.9675376890000003 podStartE2EDuration="2.967537689s" podCreationTimestamp="2025-10-01 09:51:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:51:25.96327895 +0000 UTC m=+918.078423117" watchObservedRunningTime="2025-10-01 09:51:25.967537689 +0000 UTC m=+918.082681846" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.978230 4787 scope.go:117] "RemoveContainer" containerID="0bbd6dfd786381be1dc89832debdc18140f5a5d66be20be5f53c17822216812a" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.984948 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d86d68bf7-mn47f"] Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.991753 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5d86d68bf7-mn47f"] Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.997901 4787 scope.go:117] "RemoveContainer" containerID="9312726f8da9cacc2afb441fb3742aa3356b2987cc2f23c6192836a2fa7d2e18" Oct 01 09:51:25 crc kubenswrapper[4787]: E1001 09:51:25.998404 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9312726f8da9cacc2afb441fb3742aa3356b2987cc2f23c6192836a2fa7d2e18\": container with ID starting with 9312726f8da9cacc2afb441fb3742aa3356b2987cc2f23c6192836a2fa7d2e18 not found: ID does not exist" containerID="9312726f8da9cacc2afb441fb3742aa3356b2987cc2f23c6192836a2fa7d2e18" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.998514 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9312726f8da9cacc2afb441fb3742aa3356b2987cc2f23c6192836a2fa7d2e18"} err="failed to get container status \"9312726f8da9cacc2afb441fb3742aa3356b2987cc2f23c6192836a2fa7d2e18\": rpc error: code = NotFound desc = could not find container \"9312726f8da9cacc2afb441fb3742aa3356b2987cc2f23c6192836a2fa7d2e18\": container with ID starting with 9312726f8da9cacc2afb441fb3742aa3356b2987cc2f23c6192836a2fa7d2e18 not found: ID does not exist" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.998625 4787 scope.go:117] "RemoveContainer" containerID="0bbd6dfd786381be1dc89832debdc18140f5a5d66be20be5f53c17822216812a" Oct 01 09:51:25 crc kubenswrapper[4787]: E1001 09:51:25.999046 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bbd6dfd786381be1dc89832debdc18140f5a5d66be20be5f53c17822216812a\": container with ID starting with 0bbd6dfd786381be1dc89832debdc18140f5a5d66be20be5f53c17822216812a not found: ID does not exist" containerID="0bbd6dfd786381be1dc89832debdc18140f5a5d66be20be5f53c17822216812a" Oct 01 09:51:25 crc kubenswrapper[4787]: I1001 09:51:25.999120 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bbd6dfd786381be1dc89832debdc18140f5a5d66be20be5f53c17822216812a"} err="failed to get container status \"0bbd6dfd786381be1dc89832debdc18140f5a5d66be20be5f53c17822216812a\": rpc error: code = NotFound desc = could not find container \"0bbd6dfd786381be1dc89832debdc18140f5a5d66be20be5f53c17822216812a\": container with ID starting with 0bbd6dfd786381be1dc89832debdc18140f5a5d66be20be5f53c17822216812a not found: ID does not exist" Oct 01 09:51:26 crc kubenswrapper[4787]: I1001 09:51:26.125975 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3432bc16-99b4-431a-aeb1-600e826bbc3e-etc-swift\") pod \"swift-storage-0\" (UID: \"3432bc16-99b4-431a-aeb1-600e826bbc3e\") " pod="openstack/swift-storage-0" Oct 01 09:51:26 crc kubenswrapper[4787]: E1001 09:51:26.126156 4787 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 01 09:51:26 crc kubenswrapper[4787]: E1001 09:51:26.126172 4787 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 01 09:51:26 crc kubenswrapper[4787]: E1001 09:51:26.126221 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3432bc16-99b4-431a-aeb1-600e826bbc3e-etc-swift podName:3432bc16-99b4-431a-aeb1-600e826bbc3e nodeName:}" failed. No retries permitted until 2025-10-01 09:51:28.126207535 +0000 UTC m=+920.241351692 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/3432bc16-99b4-431a-aeb1-600e826bbc3e-etc-swift") pod "swift-storage-0" (UID: "3432bc16-99b4-431a-aeb1-600e826bbc3e") : configmap "swift-ring-files" not found Oct 01 09:51:26 crc kubenswrapper[4787]: I1001 09:51:26.534472 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26ae359c-09e5-48c8-8f17-a5e6d7b75529" path="/var/lib/kubelet/pods/26ae359c-09e5-48c8-8f17-a5e6d7b75529/volumes" Oct 01 09:51:28 crc kubenswrapper[4787]: I1001 09:51:28.162799 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3432bc16-99b4-431a-aeb1-600e826bbc3e-etc-swift\") pod \"swift-storage-0\" (UID: \"3432bc16-99b4-431a-aeb1-600e826bbc3e\") " pod="openstack/swift-storage-0" Oct 01 09:51:28 crc kubenswrapper[4787]: E1001 09:51:28.163242 4787 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 01 09:51:28 crc kubenswrapper[4787]: E1001 09:51:28.163680 4787 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 01 09:51:28 crc kubenswrapper[4787]: E1001 09:51:28.163793 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3432bc16-99b4-431a-aeb1-600e826bbc3e-etc-swift podName:3432bc16-99b4-431a-aeb1-600e826bbc3e nodeName:}" failed. No retries permitted until 2025-10-01 09:51:32.163770202 +0000 UTC m=+924.278914359 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/3432bc16-99b4-431a-aeb1-600e826bbc3e-etc-swift") pod "swift-storage-0" (UID: "3432bc16-99b4-431a-aeb1-600e826bbc3e") : configmap "swift-ring-files" not found Oct 01 09:51:29 crc kubenswrapper[4787]: I1001 09:51:29.990846 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-kkdkg" event={"ID":"ed714165-98db-45a6-b96b-69601bff3e56","Type":"ContainerStarted","Data":"0bf6f70bad1ad4b963bcec190dbce79653f33154f4211677c6b43bf2fa4ed4cf"} Oct 01 09:51:29 crc kubenswrapper[4787]: I1001 09:51:29.993953 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"469f4dc6-d176-4497-ac69-6130a375737b","Type":"ContainerStarted","Data":"e88e716b091e96dfbeec718541083c5b8f296f8f98fea8a472a2917ee8c45f46"} Oct 01 09:51:29 crc kubenswrapper[4787]: I1001 09:51:29.994166 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 01 09:51:30 crc kubenswrapper[4787]: I1001 09:51:30.016534 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-kkdkg" podStartSLOduration=2.461000285 podStartE2EDuration="6.016510312s" podCreationTimestamp="2025-10-01 09:51:24 +0000 UTC" firstStartedPulling="2025-10-01 09:51:25.862874857 +0000 UTC m=+917.978019014" lastFinishedPulling="2025-10-01 09:51:29.418384874 +0000 UTC m=+921.533529041" observedRunningTime="2025-10-01 09:51:30.009359568 +0000 UTC m=+922.124503735" watchObservedRunningTime="2025-10-01 09:51:30.016510312 +0000 UTC m=+922.131654469" Oct 01 09:51:30 crc kubenswrapper[4787]: I1001 09:51:30.026210 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=16.154396679 podStartE2EDuration="48.02618986s" podCreationTimestamp="2025-10-01 09:50:42 +0000 UTC" firstStartedPulling="2025-10-01 09:50:57.539363698 +0000 UTC m=+889.654507855" lastFinishedPulling="2025-10-01 09:51:29.411156879 +0000 UTC m=+921.526301036" observedRunningTime="2025-10-01 09:51:30.023791588 +0000 UTC m=+922.138935765" watchObservedRunningTime="2025-10-01 09:51:30.02618986 +0000 UTC m=+922.141334027" Oct 01 09:51:30 crc kubenswrapper[4787]: I1001 09:51:30.788089 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-2j6gl"] Oct 01 09:51:30 crc kubenswrapper[4787]: E1001 09:51:30.788496 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26ae359c-09e5-48c8-8f17-a5e6d7b75529" containerName="dnsmasq-dns" Oct 01 09:51:30 crc kubenswrapper[4787]: I1001 09:51:30.788509 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="26ae359c-09e5-48c8-8f17-a5e6d7b75529" containerName="dnsmasq-dns" Oct 01 09:51:30 crc kubenswrapper[4787]: E1001 09:51:30.788524 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26ae359c-09e5-48c8-8f17-a5e6d7b75529" containerName="init" Oct 01 09:51:30 crc kubenswrapper[4787]: I1001 09:51:30.788531 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="26ae359c-09e5-48c8-8f17-a5e6d7b75529" containerName="init" Oct 01 09:51:30 crc kubenswrapper[4787]: I1001 09:51:30.788726 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="26ae359c-09e5-48c8-8f17-a5e6d7b75529" containerName="dnsmasq-dns" Oct 01 09:51:30 crc kubenswrapper[4787]: I1001 09:51:30.789436 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-2j6gl" Oct 01 09:51:30 crc kubenswrapper[4787]: I1001 09:51:30.799295 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-2j6gl"] Oct 01 09:51:30 crc kubenswrapper[4787]: I1001 09:51:30.817592 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfz6w\" (UniqueName: \"kubernetes.io/projected/535755f5-70d6-4fa0-8be1-c8add5e1283c-kube-api-access-gfz6w\") pod \"keystone-db-create-2j6gl\" (UID: \"535755f5-70d6-4fa0-8be1-c8add5e1283c\") " pod="openstack/keystone-db-create-2j6gl" Oct 01 09:51:30 crc kubenswrapper[4787]: I1001 09:51:30.919725 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfz6w\" (UniqueName: \"kubernetes.io/projected/535755f5-70d6-4fa0-8be1-c8add5e1283c-kube-api-access-gfz6w\") pod \"keystone-db-create-2j6gl\" (UID: \"535755f5-70d6-4fa0-8be1-c8add5e1283c\") " pod="openstack/keystone-db-create-2j6gl" Oct 01 09:51:30 crc kubenswrapper[4787]: I1001 09:51:30.939760 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfz6w\" (UniqueName: \"kubernetes.io/projected/535755f5-70d6-4fa0-8be1-c8add5e1283c-kube-api-access-gfz6w\") pod \"keystone-db-create-2j6gl\" (UID: \"535755f5-70d6-4fa0-8be1-c8add5e1283c\") " pod="openstack/keystone-db-create-2j6gl" Oct 01 09:51:31 crc kubenswrapper[4787]: I1001 09:51:31.107199 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-2j6gl" Oct 01 09:51:31 crc kubenswrapper[4787]: I1001 09:51:31.240850 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-7321-account-create-gvbhd"] Oct 01 09:51:31 crc kubenswrapper[4787]: I1001 09:51:31.243703 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7321-account-create-gvbhd" Oct 01 09:51:31 crc kubenswrapper[4787]: I1001 09:51:31.247361 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 01 09:51:31 crc kubenswrapper[4787]: I1001 09:51:31.254867 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7321-account-create-gvbhd"] Oct 01 09:51:31 crc kubenswrapper[4787]: I1001 09:51:31.328876 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jt87\" (UniqueName: \"kubernetes.io/projected/c3cbab8d-b6e2-420a-8456-efef86923af5-kube-api-access-2jt87\") pod \"placement-7321-account-create-gvbhd\" (UID: \"c3cbab8d-b6e2-420a-8456-efef86923af5\") " pod="openstack/placement-7321-account-create-gvbhd" Oct 01 09:51:31 crc kubenswrapper[4787]: I1001 09:51:31.431379 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jt87\" (UniqueName: \"kubernetes.io/projected/c3cbab8d-b6e2-420a-8456-efef86923af5-kube-api-access-2jt87\") pod \"placement-7321-account-create-gvbhd\" (UID: \"c3cbab8d-b6e2-420a-8456-efef86923af5\") " pod="openstack/placement-7321-account-create-gvbhd" Oct 01 09:51:31 crc kubenswrapper[4787]: I1001 09:51:31.464615 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jt87\" (UniqueName: \"kubernetes.io/projected/c3cbab8d-b6e2-420a-8456-efef86923af5-kube-api-access-2jt87\") pod \"placement-7321-account-create-gvbhd\" (UID: \"c3cbab8d-b6e2-420a-8456-efef86923af5\") " pod="openstack/placement-7321-account-create-gvbhd" Oct 01 09:51:31 crc kubenswrapper[4787]: I1001 09:51:31.581007 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-27c6-account-create-hqf2p"] Oct 01 09:51:31 crc kubenswrapper[4787]: I1001 09:51:31.582399 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-27c6-account-create-hqf2p" Oct 01 09:51:31 crc kubenswrapper[4787]: I1001 09:51:31.587932 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 01 09:51:31 crc kubenswrapper[4787]: I1001 09:51:31.590165 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-27c6-account-create-hqf2p"] Oct 01 09:51:31 crc kubenswrapper[4787]: I1001 09:51:31.602412 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-2j6gl"] Oct 01 09:51:31 crc kubenswrapper[4787]: I1001 09:51:31.621606 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7321-account-create-gvbhd" Oct 01 09:51:31 crc kubenswrapper[4787]: I1001 09:51:31.639227 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xwfc\" (UniqueName: \"kubernetes.io/projected/7a3b9c9c-d4f0-4f2e-b2f3-44689860ec29-kube-api-access-6xwfc\") pod \"glance-27c6-account-create-hqf2p\" (UID: \"7a3b9c9c-d4f0-4f2e-b2f3-44689860ec29\") " pod="openstack/glance-27c6-account-create-hqf2p" Oct 01 09:51:31 crc kubenswrapper[4787]: I1001 09:51:31.740936 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xwfc\" (UniqueName: \"kubernetes.io/projected/7a3b9c9c-d4f0-4f2e-b2f3-44689860ec29-kube-api-access-6xwfc\") pod \"glance-27c6-account-create-hqf2p\" (UID: \"7a3b9c9c-d4f0-4f2e-b2f3-44689860ec29\") " pod="openstack/glance-27c6-account-create-hqf2p" Oct 01 09:51:31 crc kubenswrapper[4787]: I1001 09:51:31.763920 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xwfc\" (UniqueName: \"kubernetes.io/projected/7a3b9c9c-d4f0-4f2e-b2f3-44689860ec29-kube-api-access-6xwfc\") pod \"glance-27c6-account-create-hqf2p\" (UID: \"7a3b9c9c-d4f0-4f2e-b2f3-44689860ec29\") " pod="openstack/glance-27c6-account-create-hqf2p" Oct 01 09:51:31 crc kubenswrapper[4787]: I1001 09:51:31.956304 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-27c6-account-create-hqf2p" Oct 01 09:51:32 crc kubenswrapper[4787]: I1001 09:51:32.022867 4787 generic.go:334] "Generic (PLEG): container finished" podID="ce587847-47c2-41de-95d1-c9f8ab88961e" containerID="e46335d7899902dbb16ab47d03f56851aae7b293726f79ce1a33fe0fda821f6e" exitCode=0 Oct 01 09:51:32 crc kubenswrapper[4787]: I1001 09:51:32.022923 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ce587847-47c2-41de-95d1-c9f8ab88961e","Type":"ContainerDied","Data":"e46335d7899902dbb16ab47d03f56851aae7b293726f79ce1a33fe0fda821f6e"} Oct 01 09:51:32 crc kubenswrapper[4787]: I1001 09:51:32.029596 4787 generic.go:334] "Generic (PLEG): container finished" podID="535755f5-70d6-4fa0-8be1-c8add5e1283c" containerID="9615249f72189c58f8030de6ab45f6725cd8c892d5b91b1d6092b363c976998d" exitCode=0 Oct 01 09:51:32 crc kubenswrapper[4787]: I1001 09:51:32.029663 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-2j6gl" event={"ID":"535755f5-70d6-4fa0-8be1-c8add5e1283c","Type":"ContainerDied","Data":"9615249f72189c58f8030de6ab45f6725cd8c892d5b91b1d6092b363c976998d"} Oct 01 09:51:32 crc kubenswrapper[4787]: I1001 09:51:32.029695 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-2j6gl" event={"ID":"535755f5-70d6-4fa0-8be1-c8add5e1283c","Type":"ContainerStarted","Data":"99dd4de56c65906fca3be526ccb2d53df35978f7cbbdf83b88ce18f9fab84959"} Oct 01 09:51:32 crc kubenswrapper[4787]: I1001 09:51:32.036005 4787 generic.go:334] "Generic (PLEG): container finished" podID="e762572f-d5b9-462e-9953-7143f648c9ae" containerID="2c2684e2732e0784cacbce917a9aa23a444a150b54d2e11ce18ac1efd3f00bd0" exitCode=0 Oct 01 09:51:32 crc kubenswrapper[4787]: I1001 09:51:32.036058 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e762572f-d5b9-462e-9953-7143f648c9ae","Type":"ContainerDied","Data":"2c2684e2732e0784cacbce917a9aa23a444a150b54d2e11ce18ac1efd3f00bd0"} Oct 01 09:51:32 crc kubenswrapper[4787]: I1001 09:51:32.083739 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7321-account-create-gvbhd"] Oct 01 09:51:32 crc kubenswrapper[4787]: I1001 09:51:32.256889 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3432bc16-99b4-431a-aeb1-600e826bbc3e-etc-swift\") pod \"swift-storage-0\" (UID: \"3432bc16-99b4-431a-aeb1-600e826bbc3e\") " pod="openstack/swift-storage-0" Oct 01 09:51:32 crc kubenswrapper[4787]: E1001 09:51:32.257094 4787 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 01 09:51:32 crc kubenswrapper[4787]: E1001 09:51:32.257107 4787 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 01 09:51:32 crc kubenswrapper[4787]: E1001 09:51:32.257148 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3432bc16-99b4-431a-aeb1-600e826bbc3e-etc-swift podName:3432bc16-99b4-431a-aeb1-600e826bbc3e nodeName:}" failed. No retries permitted until 2025-10-01 09:51:40.257135681 +0000 UTC m=+932.372279838 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/3432bc16-99b4-431a-aeb1-600e826bbc3e-etc-swift") pod "swift-storage-0" (UID: "3432bc16-99b4-431a-aeb1-600e826bbc3e") : configmap "swift-ring-files" not found Oct 01 09:51:32 crc kubenswrapper[4787]: I1001 09:51:32.429838 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-27c6-account-create-hqf2p"] Oct 01 09:51:32 crc kubenswrapper[4787]: W1001 09:51:32.436875 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7a3b9c9c_d4f0_4f2e_b2f3_44689860ec29.slice/crio-c913339cf56aa14c9693270c51ea60b26361e76d03918488ddedcb778cf5ef9e WatchSource:0}: Error finding container c913339cf56aa14c9693270c51ea60b26361e76d03918488ddedcb778cf5ef9e: Status 404 returned error can't find the container with id c913339cf56aa14c9693270c51ea60b26361e76d03918488ddedcb778cf5ef9e Oct 01 09:51:33 crc kubenswrapper[4787]: I1001 09:51:33.049315 4787 generic.go:334] "Generic (PLEG): container finished" podID="7a3b9c9c-d4f0-4f2e-b2f3-44689860ec29" containerID="65b2293dbb8338d9092baa098fc300c3527196b0d2d7eaaf08fdb65b4dea50b7" exitCode=0 Oct 01 09:51:33 crc kubenswrapper[4787]: I1001 09:51:33.049417 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-27c6-account-create-hqf2p" event={"ID":"7a3b9c9c-d4f0-4f2e-b2f3-44689860ec29","Type":"ContainerDied","Data":"65b2293dbb8338d9092baa098fc300c3527196b0d2d7eaaf08fdb65b4dea50b7"} Oct 01 09:51:33 crc kubenswrapper[4787]: I1001 09:51:33.049470 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-27c6-account-create-hqf2p" event={"ID":"7a3b9c9c-d4f0-4f2e-b2f3-44689860ec29","Type":"ContainerStarted","Data":"c913339cf56aa14c9693270c51ea60b26361e76d03918488ddedcb778cf5ef9e"} Oct 01 09:51:33 crc kubenswrapper[4787]: I1001 09:51:33.051700 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e762572f-d5b9-462e-9953-7143f648c9ae","Type":"ContainerStarted","Data":"dc3a7667b021a77fbd7004c14e7eea8bcaffddf0b1788a0efc6e5c22d34ef255"} Oct 01 09:51:33 crc kubenswrapper[4787]: I1001 09:51:33.052187 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 01 09:51:33 crc kubenswrapper[4787]: I1001 09:51:33.055709 4787 generic.go:334] "Generic (PLEG): container finished" podID="c3cbab8d-b6e2-420a-8456-efef86923af5" containerID="dfc015bb702961c5fad27e0d1f72d6de0ae82cd36b5eece674ed64f963d1e937" exitCode=0 Oct 01 09:51:33 crc kubenswrapper[4787]: I1001 09:51:33.055768 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7321-account-create-gvbhd" event={"ID":"c3cbab8d-b6e2-420a-8456-efef86923af5","Type":"ContainerDied","Data":"dfc015bb702961c5fad27e0d1f72d6de0ae82cd36b5eece674ed64f963d1e937"} Oct 01 09:51:33 crc kubenswrapper[4787]: I1001 09:51:33.055785 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7321-account-create-gvbhd" event={"ID":"c3cbab8d-b6e2-420a-8456-efef86923af5","Type":"ContainerStarted","Data":"612a735b8a393224c6484d069c1f04e19f5cdd4c82621a1710c3ffaec20450a5"} Oct 01 09:51:33 crc kubenswrapper[4787]: I1001 09:51:33.059223 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ce587847-47c2-41de-95d1-c9f8ab88961e","Type":"ContainerStarted","Data":"7fc382e561b660b35dc13b61c6cea44e5887caaef5da418260ae272e6cffb258"} Oct 01 09:51:33 crc kubenswrapper[4787]: I1001 09:51:33.059621 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:51:33 crc kubenswrapper[4787]: I1001 09:51:33.102585 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.321293515 podStartE2EDuration="57.102568217s" podCreationTimestamp="2025-10-01 09:50:36 +0000 UTC" firstStartedPulling="2025-10-01 09:50:38.849374006 +0000 UTC m=+870.964518163" lastFinishedPulling="2025-10-01 09:50:57.630648708 +0000 UTC m=+889.745792865" observedRunningTime="2025-10-01 09:51:33.097726513 +0000 UTC m=+925.212870680" watchObservedRunningTime="2025-10-01 09:51:33.102568217 +0000 UTC m=+925.217712374" Oct 01 09:51:33 crc kubenswrapper[4787]: I1001 09:51:33.128820 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.941228855 podStartE2EDuration="57.128794249s" podCreationTimestamp="2025-10-01 09:50:36 +0000 UTC" firstStartedPulling="2025-10-01 09:50:38.442862788 +0000 UTC m=+870.558006945" lastFinishedPulling="2025-10-01 09:50:57.630428182 +0000 UTC m=+889.745572339" observedRunningTime="2025-10-01 09:51:33.118154506 +0000 UTC m=+925.233298673" watchObservedRunningTime="2025-10-01 09:51:33.128794249 +0000 UTC m=+925.243938406" Oct 01 09:51:33 crc kubenswrapper[4787]: I1001 09:51:33.429718 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-2j6gl" Oct 01 09:51:33 crc kubenswrapper[4787]: I1001 09:51:33.582050 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfz6w\" (UniqueName: \"kubernetes.io/projected/535755f5-70d6-4fa0-8be1-c8add5e1283c-kube-api-access-gfz6w\") pod \"535755f5-70d6-4fa0-8be1-c8add5e1283c\" (UID: \"535755f5-70d6-4fa0-8be1-c8add5e1283c\") " Oct 01 09:51:33 crc kubenswrapper[4787]: I1001 09:51:33.588599 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/535755f5-70d6-4fa0-8be1-c8add5e1283c-kube-api-access-gfz6w" (OuterVolumeSpecName: "kube-api-access-gfz6w") pod "535755f5-70d6-4fa0-8be1-c8add5e1283c" (UID: "535755f5-70d6-4fa0-8be1-c8add5e1283c"). InnerVolumeSpecName "kube-api-access-gfz6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:51:33 crc kubenswrapper[4787]: I1001 09:51:33.683859 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfz6w\" (UniqueName: \"kubernetes.io/projected/535755f5-70d6-4fa0-8be1-c8add5e1283c-kube-api-access-gfz6w\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:33 crc kubenswrapper[4787]: I1001 09:51:33.757221 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6c6d5d5bd7-4mbvj" Oct 01 09:51:33 crc kubenswrapper[4787]: I1001 09:51:33.813784 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77795d58f5-kh2gz"] Oct 01 09:51:33 crc kubenswrapper[4787]: I1001 09:51:33.814158 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77795d58f5-kh2gz" podUID="105b8088-a178-492b-aefd-20a5e6ad0dc8" containerName="dnsmasq-dns" containerID="cri-o://fd066a586c077d34157337d39d10e12ad72554813b7b93ab4f4c3567d5ef86c7" gracePeriod=10 Oct 01 09:51:34 crc kubenswrapper[4787]: I1001 09:51:34.104704 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-2j6gl" Oct 01 09:51:34 crc kubenswrapper[4787]: I1001 09:51:34.105363 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-2j6gl" event={"ID":"535755f5-70d6-4fa0-8be1-c8add5e1283c","Type":"ContainerDied","Data":"99dd4de56c65906fca3be526ccb2d53df35978f7cbbdf83b88ce18f9fab84959"} Oct 01 09:51:34 crc kubenswrapper[4787]: I1001 09:51:34.105441 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="99dd4de56c65906fca3be526ccb2d53df35978f7cbbdf83b88ce18f9fab84959" Oct 01 09:51:34 crc kubenswrapper[4787]: I1001 09:51:34.127457 4787 generic.go:334] "Generic (PLEG): container finished" podID="105b8088-a178-492b-aefd-20a5e6ad0dc8" containerID="fd066a586c077d34157337d39d10e12ad72554813b7b93ab4f4c3567d5ef86c7" exitCode=0 Oct 01 09:51:34 crc kubenswrapper[4787]: I1001 09:51:34.127548 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77795d58f5-kh2gz" event={"ID":"105b8088-a178-492b-aefd-20a5e6ad0dc8","Type":"ContainerDied","Data":"fd066a586c077d34157337d39d10e12ad72554813b7b93ab4f4c3567d5ef86c7"} Oct 01 09:51:34 crc kubenswrapper[4787]: I1001 09:51:34.418436 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77795d58f5-kh2gz" Oct 01 09:51:34 crc kubenswrapper[4787]: I1001 09:51:34.497641 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/105b8088-a178-492b-aefd-20a5e6ad0dc8-dns-svc\") pod \"105b8088-a178-492b-aefd-20a5e6ad0dc8\" (UID: \"105b8088-a178-492b-aefd-20a5e6ad0dc8\") " Oct 01 09:51:34 crc kubenswrapper[4787]: I1001 09:51:34.497704 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnjkn\" (UniqueName: \"kubernetes.io/projected/105b8088-a178-492b-aefd-20a5e6ad0dc8-kube-api-access-rnjkn\") pod \"105b8088-a178-492b-aefd-20a5e6ad0dc8\" (UID: \"105b8088-a178-492b-aefd-20a5e6ad0dc8\") " Oct 01 09:51:34 crc kubenswrapper[4787]: I1001 09:51:34.497757 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/105b8088-a178-492b-aefd-20a5e6ad0dc8-config\") pod \"105b8088-a178-492b-aefd-20a5e6ad0dc8\" (UID: \"105b8088-a178-492b-aefd-20a5e6ad0dc8\") " Oct 01 09:51:34 crc kubenswrapper[4787]: I1001 09:51:34.521371 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/105b8088-a178-492b-aefd-20a5e6ad0dc8-kube-api-access-rnjkn" (OuterVolumeSpecName: "kube-api-access-rnjkn") pod "105b8088-a178-492b-aefd-20a5e6ad0dc8" (UID: "105b8088-a178-492b-aefd-20a5e6ad0dc8"). InnerVolumeSpecName "kube-api-access-rnjkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:51:34 crc kubenswrapper[4787]: I1001 09:51:34.585315 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/105b8088-a178-492b-aefd-20a5e6ad0dc8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "105b8088-a178-492b-aefd-20a5e6ad0dc8" (UID: "105b8088-a178-492b-aefd-20a5e6ad0dc8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:51:34 crc kubenswrapper[4787]: I1001 09:51:34.599238 4787 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/105b8088-a178-492b-aefd-20a5e6ad0dc8-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:34 crc kubenswrapper[4787]: I1001 09:51:34.599279 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnjkn\" (UniqueName: \"kubernetes.io/projected/105b8088-a178-492b-aefd-20a5e6ad0dc8-kube-api-access-rnjkn\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:34 crc kubenswrapper[4787]: I1001 09:51:34.636902 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/105b8088-a178-492b-aefd-20a5e6ad0dc8-config" (OuterVolumeSpecName: "config") pod "105b8088-a178-492b-aefd-20a5e6ad0dc8" (UID: "105b8088-a178-492b-aefd-20a5e6ad0dc8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:51:34 crc kubenswrapper[4787]: I1001 09:51:34.648751 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7321-account-create-gvbhd" Oct 01 09:51:34 crc kubenswrapper[4787]: I1001 09:51:34.652659 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-27c6-account-create-hqf2p" Oct 01 09:51:34 crc kubenswrapper[4787]: I1001 09:51:34.700835 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jt87\" (UniqueName: \"kubernetes.io/projected/c3cbab8d-b6e2-420a-8456-efef86923af5-kube-api-access-2jt87\") pod \"c3cbab8d-b6e2-420a-8456-efef86923af5\" (UID: \"c3cbab8d-b6e2-420a-8456-efef86923af5\") " Oct 01 09:51:34 crc kubenswrapper[4787]: I1001 09:51:34.701005 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xwfc\" (UniqueName: \"kubernetes.io/projected/7a3b9c9c-d4f0-4f2e-b2f3-44689860ec29-kube-api-access-6xwfc\") pod \"7a3b9c9c-d4f0-4f2e-b2f3-44689860ec29\" (UID: \"7a3b9c9c-d4f0-4f2e-b2f3-44689860ec29\") " Oct 01 09:51:34 crc kubenswrapper[4787]: I1001 09:51:34.701365 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/105b8088-a178-492b-aefd-20a5e6ad0dc8-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:34 crc kubenswrapper[4787]: I1001 09:51:34.706751 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a3b9c9c-d4f0-4f2e-b2f3-44689860ec29-kube-api-access-6xwfc" (OuterVolumeSpecName: "kube-api-access-6xwfc") pod "7a3b9c9c-d4f0-4f2e-b2f3-44689860ec29" (UID: "7a3b9c9c-d4f0-4f2e-b2f3-44689860ec29"). InnerVolumeSpecName "kube-api-access-6xwfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:51:34 crc kubenswrapper[4787]: I1001 09:51:34.706858 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3cbab8d-b6e2-420a-8456-efef86923af5-kube-api-access-2jt87" (OuterVolumeSpecName: "kube-api-access-2jt87") pod "c3cbab8d-b6e2-420a-8456-efef86923af5" (UID: "c3cbab8d-b6e2-420a-8456-efef86923af5"). InnerVolumeSpecName "kube-api-access-2jt87". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:51:34 crc kubenswrapper[4787]: I1001 09:51:34.802969 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jt87\" (UniqueName: \"kubernetes.io/projected/c3cbab8d-b6e2-420a-8456-efef86923af5-kube-api-access-2jt87\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:34 crc kubenswrapper[4787]: I1001 09:51:34.803007 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xwfc\" (UniqueName: \"kubernetes.io/projected/7a3b9c9c-d4f0-4f2e-b2f3-44689860ec29-kube-api-access-6xwfc\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:35 crc kubenswrapper[4787]: I1001 09:51:35.138105 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77795d58f5-kh2gz" event={"ID":"105b8088-a178-492b-aefd-20a5e6ad0dc8","Type":"ContainerDied","Data":"f8ab9a21de90c12090d0a833ff82757b32680470016b00c2d78164bca607be41"} Oct 01 09:51:35 crc kubenswrapper[4787]: I1001 09:51:35.138263 4787 scope.go:117] "RemoveContainer" containerID="fd066a586c077d34157337d39d10e12ad72554813b7b93ab4f4c3567d5ef86c7" Oct 01 09:51:35 crc kubenswrapper[4787]: I1001 09:51:35.138478 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77795d58f5-kh2gz" Oct 01 09:51:35 crc kubenswrapper[4787]: I1001 09:51:35.139880 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7321-account-create-gvbhd" event={"ID":"c3cbab8d-b6e2-420a-8456-efef86923af5","Type":"ContainerDied","Data":"612a735b8a393224c6484d069c1f04e19f5cdd4c82621a1710c3ffaec20450a5"} Oct 01 09:51:35 crc kubenswrapper[4787]: I1001 09:51:35.139920 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="612a735b8a393224c6484d069c1f04e19f5cdd4c82621a1710c3ffaec20450a5" Oct 01 09:51:35 crc kubenswrapper[4787]: I1001 09:51:35.139971 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7321-account-create-gvbhd" Oct 01 09:51:35 crc kubenswrapper[4787]: I1001 09:51:35.142088 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-27c6-account-create-hqf2p" event={"ID":"7a3b9c9c-d4f0-4f2e-b2f3-44689860ec29","Type":"ContainerDied","Data":"c913339cf56aa14c9693270c51ea60b26361e76d03918488ddedcb778cf5ef9e"} Oct 01 09:51:35 crc kubenswrapper[4787]: I1001 09:51:35.142127 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-27c6-account-create-hqf2p" Oct 01 09:51:35 crc kubenswrapper[4787]: I1001 09:51:35.142140 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c913339cf56aa14c9693270c51ea60b26361e76d03918488ddedcb778cf5ef9e" Oct 01 09:51:35 crc kubenswrapper[4787]: I1001 09:51:35.161531 4787 scope.go:117] "RemoveContainer" containerID="ae936474381271c575e1a56f21612c7a626e640004485f0662ae8025a0ac686f" Oct 01 09:51:35 crc kubenswrapper[4787]: I1001 09:51:35.187626 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77795d58f5-kh2gz"] Oct 01 09:51:35 crc kubenswrapper[4787]: I1001 09:51:35.200797 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77795d58f5-kh2gz"] Oct 01 09:51:36 crc kubenswrapper[4787]: I1001 09:51:36.534328 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="105b8088-a178-492b-aefd-20a5e6ad0dc8" path="/var/lib/kubelet/pods/105b8088-a178-492b-aefd-20a5e6ad0dc8/volumes" Oct 01 09:51:36 crc kubenswrapper[4787]: I1001 09:51:36.716060 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-kckrj"] Oct 01 09:51:36 crc kubenswrapper[4787]: E1001 09:51:36.716811 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3cbab8d-b6e2-420a-8456-efef86923af5" containerName="mariadb-account-create" Oct 01 09:51:36 crc kubenswrapper[4787]: I1001 09:51:36.716829 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3cbab8d-b6e2-420a-8456-efef86923af5" containerName="mariadb-account-create" Oct 01 09:51:36 crc kubenswrapper[4787]: E1001 09:51:36.716854 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="105b8088-a178-492b-aefd-20a5e6ad0dc8" containerName="init" Oct 01 09:51:36 crc kubenswrapper[4787]: I1001 09:51:36.716863 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="105b8088-a178-492b-aefd-20a5e6ad0dc8" containerName="init" Oct 01 09:51:36 crc kubenswrapper[4787]: E1001 09:51:36.716874 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a3b9c9c-d4f0-4f2e-b2f3-44689860ec29" containerName="mariadb-account-create" Oct 01 09:51:36 crc kubenswrapper[4787]: I1001 09:51:36.716881 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a3b9c9c-d4f0-4f2e-b2f3-44689860ec29" containerName="mariadb-account-create" Oct 01 09:51:36 crc kubenswrapper[4787]: E1001 09:51:36.716889 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="535755f5-70d6-4fa0-8be1-c8add5e1283c" containerName="mariadb-database-create" Oct 01 09:51:36 crc kubenswrapper[4787]: I1001 09:51:36.716911 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="535755f5-70d6-4fa0-8be1-c8add5e1283c" containerName="mariadb-database-create" Oct 01 09:51:36 crc kubenswrapper[4787]: E1001 09:51:36.716921 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="105b8088-a178-492b-aefd-20a5e6ad0dc8" containerName="dnsmasq-dns" Oct 01 09:51:36 crc kubenswrapper[4787]: I1001 09:51:36.716927 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="105b8088-a178-492b-aefd-20a5e6ad0dc8" containerName="dnsmasq-dns" Oct 01 09:51:36 crc kubenswrapper[4787]: I1001 09:51:36.717093 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="105b8088-a178-492b-aefd-20a5e6ad0dc8" containerName="dnsmasq-dns" Oct 01 09:51:36 crc kubenswrapper[4787]: I1001 09:51:36.717103 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="535755f5-70d6-4fa0-8be1-c8add5e1283c" containerName="mariadb-database-create" Oct 01 09:51:36 crc kubenswrapper[4787]: I1001 09:51:36.717114 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3cbab8d-b6e2-420a-8456-efef86923af5" containerName="mariadb-account-create" Oct 01 09:51:36 crc kubenswrapper[4787]: I1001 09:51:36.717132 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a3b9c9c-d4f0-4f2e-b2f3-44689860ec29" containerName="mariadb-account-create" Oct 01 09:51:36 crc kubenswrapper[4787]: I1001 09:51:36.717656 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-kckrj" Oct 01 09:51:36 crc kubenswrapper[4787]: I1001 09:51:36.723259 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 01 09:51:36 crc kubenswrapper[4787]: I1001 09:51:36.723260 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-bskck" Oct 01 09:51:36 crc kubenswrapper[4787]: I1001 09:51:36.732875 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-kckrj"] Oct 01 09:51:36 crc kubenswrapper[4787]: I1001 09:51:36.741778 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f7251286-7de7-48ef-a418-6d67749e40b8-db-sync-config-data\") pod \"glance-db-sync-kckrj\" (UID: \"f7251286-7de7-48ef-a418-6d67749e40b8\") " pod="openstack/glance-db-sync-kckrj" Oct 01 09:51:36 crc kubenswrapper[4787]: I1001 09:51:36.741924 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2ff5\" (UniqueName: \"kubernetes.io/projected/f7251286-7de7-48ef-a418-6d67749e40b8-kube-api-access-d2ff5\") pod \"glance-db-sync-kckrj\" (UID: \"f7251286-7de7-48ef-a418-6d67749e40b8\") " pod="openstack/glance-db-sync-kckrj" Oct 01 09:51:36 crc kubenswrapper[4787]: I1001 09:51:36.741979 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7251286-7de7-48ef-a418-6d67749e40b8-config-data\") pod \"glance-db-sync-kckrj\" (UID: \"f7251286-7de7-48ef-a418-6d67749e40b8\") " pod="openstack/glance-db-sync-kckrj" Oct 01 09:51:36 crc kubenswrapper[4787]: I1001 09:51:36.742010 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7251286-7de7-48ef-a418-6d67749e40b8-combined-ca-bundle\") pod \"glance-db-sync-kckrj\" (UID: \"f7251286-7de7-48ef-a418-6d67749e40b8\") " pod="openstack/glance-db-sync-kckrj" Oct 01 09:51:36 crc kubenswrapper[4787]: I1001 09:51:36.843996 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7251286-7de7-48ef-a418-6d67749e40b8-combined-ca-bundle\") pod \"glance-db-sync-kckrj\" (UID: \"f7251286-7de7-48ef-a418-6d67749e40b8\") " pod="openstack/glance-db-sync-kckrj" Oct 01 09:51:36 crc kubenswrapper[4787]: I1001 09:51:36.844101 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f7251286-7de7-48ef-a418-6d67749e40b8-db-sync-config-data\") pod \"glance-db-sync-kckrj\" (UID: \"f7251286-7de7-48ef-a418-6d67749e40b8\") " pod="openstack/glance-db-sync-kckrj" Oct 01 09:51:36 crc kubenswrapper[4787]: I1001 09:51:36.844157 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2ff5\" (UniqueName: \"kubernetes.io/projected/f7251286-7de7-48ef-a418-6d67749e40b8-kube-api-access-d2ff5\") pod \"glance-db-sync-kckrj\" (UID: \"f7251286-7de7-48ef-a418-6d67749e40b8\") " pod="openstack/glance-db-sync-kckrj" Oct 01 09:51:36 crc kubenswrapper[4787]: I1001 09:51:36.844199 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7251286-7de7-48ef-a418-6d67749e40b8-config-data\") pod \"glance-db-sync-kckrj\" (UID: \"f7251286-7de7-48ef-a418-6d67749e40b8\") " pod="openstack/glance-db-sync-kckrj" Oct 01 09:51:36 crc kubenswrapper[4787]: I1001 09:51:36.848987 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f7251286-7de7-48ef-a418-6d67749e40b8-db-sync-config-data\") pod \"glance-db-sync-kckrj\" (UID: \"f7251286-7de7-48ef-a418-6d67749e40b8\") " pod="openstack/glance-db-sync-kckrj" Oct 01 09:51:36 crc kubenswrapper[4787]: I1001 09:51:36.851368 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7251286-7de7-48ef-a418-6d67749e40b8-config-data\") pod \"glance-db-sync-kckrj\" (UID: \"f7251286-7de7-48ef-a418-6d67749e40b8\") " pod="openstack/glance-db-sync-kckrj" Oct 01 09:51:36 crc kubenswrapper[4787]: I1001 09:51:36.851434 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7251286-7de7-48ef-a418-6d67749e40b8-combined-ca-bundle\") pod \"glance-db-sync-kckrj\" (UID: \"f7251286-7de7-48ef-a418-6d67749e40b8\") " pod="openstack/glance-db-sync-kckrj" Oct 01 09:51:36 crc kubenswrapper[4787]: I1001 09:51:36.862409 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2ff5\" (UniqueName: \"kubernetes.io/projected/f7251286-7de7-48ef-a418-6d67749e40b8-kube-api-access-d2ff5\") pod \"glance-db-sync-kckrj\" (UID: \"f7251286-7de7-48ef-a418-6d67749e40b8\") " pod="openstack/glance-db-sync-kckrj" Oct 01 09:51:36 crc kubenswrapper[4787]: I1001 09:51:36.959345 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 01 09:51:37 crc kubenswrapper[4787]: I1001 09:51:37.038300 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-kckrj" Oct 01 09:51:37 crc kubenswrapper[4787]: I1001 09:51:37.171791 4787 generic.go:334] "Generic (PLEG): container finished" podID="ed714165-98db-45a6-b96b-69601bff3e56" containerID="0bf6f70bad1ad4b963bcec190dbce79653f33154f4211677c6b43bf2fa4ed4cf" exitCode=0 Oct 01 09:51:37 crc kubenswrapper[4787]: I1001 09:51:37.171869 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-kkdkg" event={"ID":"ed714165-98db-45a6-b96b-69601bff3e56","Type":"ContainerDied","Data":"0bf6f70bad1ad4b963bcec190dbce79653f33154f4211677c6b43bf2fa4ed4cf"} Oct 01 09:51:37 crc kubenswrapper[4787]: I1001 09:51:37.676545 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-kckrj"] Oct 01 09:51:38 crc kubenswrapper[4787]: I1001 09:51:38.180015 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-kckrj" event={"ID":"f7251286-7de7-48ef-a418-6d67749e40b8","Type":"ContainerStarted","Data":"09a402eafdfd770a947714471641ca787f0259aa361fa442756a49dfbd67ef7f"} Oct 01 09:51:38 crc kubenswrapper[4787]: I1001 09:51:38.498892 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-kkdkg" Oct 01 09:51:38 crc kubenswrapper[4787]: I1001 09:51:38.579889 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ed714165-98db-45a6-b96b-69601bff3e56-scripts\") pod \"ed714165-98db-45a6-b96b-69601bff3e56\" (UID: \"ed714165-98db-45a6-b96b-69601bff3e56\") " Oct 01 09:51:38 crc kubenswrapper[4787]: I1001 09:51:38.580042 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed714165-98db-45a6-b96b-69601bff3e56-combined-ca-bundle\") pod \"ed714165-98db-45a6-b96b-69601bff3e56\" (UID: \"ed714165-98db-45a6-b96b-69601bff3e56\") " Oct 01 09:51:38 crc kubenswrapper[4787]: I1001 09:51:38.580147 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ed714165-98db-45a6-b96b-69601bff3e56-swiftconf\") pod \"ed714165-98db-45a6-b96b-69601bff3e56\" (UID: \"ed714165-98db-45a6-b96b-69601bff3e56\") " Oct 01 09:51:38 crc kubenswrapper[4787]: I1001 09:51:38.580209 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ed714165-98db-45a6-b96b-69601bff3e56-dispersionconf\") pod \"ed714165-98db-45a6-b96b-69601bff3e56\" (UID: \"ed714165-98db-45a6-b96b-69601bff3e56\") " Oct 01 09:51:38 crc kubenswrapper[4787]: I1001 09:51:38.580262 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ed714165-98db-45a6-b96b-69601bff3e56-ring-data-devices\") pod \"ed714165-98db-45a6-b96b-69601bff3e56\" (UID: \"ed714165-98db-45a6-b96b-69601bff3e56\") " Oct 01 09:51:38 crc kubenswrapper[4787]: I1001 09:51:38.580362 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqg4v\" (UniqueName: \"kubernetes.io/projected/ed714165-98db-45a6-b96b-69601bff3e56-kube-api-access-sqg4v\") pod \"ed714165-98db-45a6-b96b-69601bff3e56\" (UID: \"ed714165-98db-45a6-b96b-69601bff3e56\") " Oct 01 09:51:38 crc kubenswrapper[4787]: I1001 09:51:38.580432 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ed714165-98db-45a6-b96b-69601bff3e56-etc-swift\") pod \"ed714165-98db-45a6-b96b-69601bff3e56\" (UID: \"ed714165-98db-45a6-b96b-69601bff3e56\") " Oct 01 09:51:38 crc kubenswrapper[4787]: I1001 09:51:38.582555 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed714165-98db-45a6-b96b-69601bff3e56-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "ed714165-98db-45a6-b96b-69601bff3e56" (UID: "ed714165-98db-45a6-b96b-69601bff3e56"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:51:38 crc kubenswrapper[4787]: I1001 09:51:38.583181 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed714165-98db-45a6-b96b-69601bff3e56-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "ed714165-98db-45a6-b96b-69601bff3e56" (UID: "ed714165-98db-45a6-b96b-69601bff3e56"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:51:38 crc kubenswrapper[4787]: I1001 09:51:38.592802 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed714165-98db-45a6-b96b-69601bff3e56-kube-api-access-sqg4v" (OuterVolumeSpecName: "kube-api-access-sqg4v") pod "ed714165-98db-45a6-b96b-69601bff3e56" (UID: "ed714165-98db-45a6-b96b-69601bff3e56"). InnerVolumeSpecName "kube-api-access-sqg4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:51:38 crc kubenswrapper[4787]: I1001 09:51:38.607941 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed714165-98db-45a6-b96b-69601bff3e56-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "ed714165-98db-45a6-b96b-69601bff3e56" (UID: "ed714165-98db-45a6-b96b-69601bff3e56"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:51:38 crc kubenswrapper[4787]: I1001 09:51:38.617504 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed714165-98db-45a6-b96b-69601bff3e56-scripts" (OuterVolumeSpecName: "scripts") pod "ed714165-98db-45a6-b96b-69601bff3e56" (UID: "ed714165-98db-45a6-b96b-69601bff3e56"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:51:38 crc kubenswrapper[4787]: I1001 09:51:38.617486 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed714165-98db-45a6-b96b-69601bff3e56-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ed714165-98db-45a6-b96b-69601bff3e56" (UID: "ed714165-98db-45a6-b96b-69601bff3e56"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:51:38 crc kubenswrapper[4787]: I1001 09:51:38.633553 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed714165-98db-45a6-b96b-69601bff3e56-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "ed714165-98db-45a6-b96b-69601bff3e56" (UID: "ed714165-98db-45a6-b96b-69601bff3e56"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:51:38 crc kubenswrapper[4787]: I1001 09:51:38.682092 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqg4v\" (UniqueName: \"kubernetes.io/projected/ed714165-98db-45a6-b96b-69601bff3e56-kube-api-access-sqg4v\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:38 crc kubenswrapper[4787]: I1001 09:51:38.682124 4787 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ed714165-98db-45a6-b96b-69601bff3e56-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:38 crc kubenswrapper[4787]: I1001 09:51:38.682136 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ed714165-98db-45a6-b96b-69601bff3e56-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:38 crc kubenswrapper[4787]: I1001 09:51:38.682146 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed714165-98db-45a6-b96b-69601bff3e56-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:38 crc kubenswrapper[4787]: I1001 09:51:38.682156 4787 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ed714165-98db-45a6-b96b-69601bff3e56-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:38 crc kubenswrapper[4787]: I1001 09:51:38.682164 4787 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ed714165-98db-45a6-b96b-69601bff3e56-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:38 crc kubenswrapper[4787]: I1001 09:51:38.682173 4787 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ed714165-98db-45a6-b96b-69601bff3e56-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:39 crc kubenswrapper[4787]: I1001 09:51:39.203932 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-kkdkg" event={"ID":"ed714165-98db-45a6-b96b-69601bff3e56","Type":"ContainerDied","Data":"298e605666808eb9c6180bfe0fb573ee731369748b6db8ea79a1150ee8ec384e"} Oct 01 09:51:39 crc kubenswrapper[4787]: I1001 09:51:39.204374 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="298e605666808eb9c6180bfe0fb573ee731369748b6db8ea79a1150ee8ec384e" Oct 01 09:51:39 crc kubenswrapper[4787]: I1001 09:51:39.204229 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-kkdkg" Oct 01 09:51:40 crc kubenswrapper[4787]: I1001 09:51:40.304488 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3432bc16-99b4-431a-aeb1-600e826bbc3e-etc-swift\") pod \"swift-storage-0\" (UID: \"3432bc16-99b4-431a-aeb1-600e826bbc3e\") " pod="openstack/swift-storage-0" Oct 01 09:51:40 crc kubenswrapper[4787]: I1001 09:51:40.316965 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3432bc16-99b4-431a-aeb1-600e826bbc3e-etc-swift\") pod \"swift-storage-0\" (UID: \"3432bc16-99b4-431a-aeb1-600e826bbc3e\") " pod="openstack/swift-storage-0" Oct 01 09:51:40 crc kubenswrapper[4787]: I1001 09:51:40.372982 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 01 09:51:40 crc kubenswrapper[4787]: I1001 09:51:40.914211 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-de61-account-create-hxhr8"] Oct 01 09:51:40 crc kubenswrapper[4787]: E1001 09:51:40.914929 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed714165-98db-45a6-b96b-69601bff3e56" containerName="swift-ring-rebalance" Oct 01 09:51:40 crc kubenswrapper[4787]: I1001 09:51:40.914955 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed714165-98db-45a6-b96b-69601bff3e56" containerName="swift-ring-rebalance" Oct 01 09:51:40 crc kubenswrapper[4787]: I1001 09:51:40.915169 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed714165-98db-45a6-b96b-69601bff3e56" containerName="swift-ring-rebalance" Oct 01 09:51:40 crc kubenswrapper[4787]: I1001 09:51:40.915894 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-de61-account-create-hxhr8" Oct 01 09:51:40 crc kubenswrapper[4787]: I1001 09:51:40.923992 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-de61-account-create-hxhr8"] Oct 01 09:51:40 crc kubenswrapper[4787]: I1001 09:51:40.929870 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 01 09:51:40 crc kubenswrapper[4787]: I1001 09:51:40.987973 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 01 09:51:40 crc kubenswrapper[4787]: W1001 09:51:40.999814 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3432bc16_99b4_431a_aeb1_600e826bbc3e.slice/crio-48459f49420cf39e53e718b702999bb29b7d4b0b4236fc3808fd7ff3b87d2c55 WatchSource:0}: Error finding container 48459f49420cf39e53e718b702999bb29b7d4b0b4236fc3808fd7ff3b87d2c55: Status 404 returned error can't find the container with id 48459f49420cf39e53e718b702999bb29b7d4b0b4236fc3808fd7ff3b87d2c55 Oct 01 09:51:41 crc kubenswrapper[4787]: I1001 09:51:41.032975 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8b2t\" (UniqueName: \"kubernetes.io/projected/1d2457dd-a699-48b4-86d0-69505da5345c-kube-api-access-g8b2t\") pod \"keystone-de61-account-create-hxhr8\" (UID: \"1d2457dd-a699-48b4-86d0-69505da5345c\") " pod="openstack/keystone-de61-account-create-hxhr8" Oct 01 09:51:41 crc kubenswrapper[4787]: I1001 09:51:41.134469 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8b2t\" (UniqueName: \"kubernetes.io/projected/1d2457dd-a699-48b4-86d0-69505da5345c-kube-api-access-g8b2t\") pod \"keystone-de61-account-create-hxhr8\" (UID: \"1d2457dd-a699-48b4-86d0-69505da5345c\") " pod="openstack/keystone-de61-account-create-hxhr8" Oct 01 09:51:41 crc kubenswrapper[4787]: I1001 09:51:41.155856 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8b2t\" (UniqueName: \"kubernetes.io/projected/1d2457dd-a699-48b4-86d0-69505da5345c-kube-api-access-g8b2t\") pod \"keystone-de61-account-create-hxhr8\" (UID: \"1d2457dd-a699-48b4-86d0-69505da5345c\") " pod="openstack/keystone-de61-account-create-hxhr8" Oct 01 09:51:41 crc kubenswrapper[4787]: I1001 09:51:41.221499 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3432bc16-99b4-431a-aeb1-600e826bbc3e","Type":"ContainerStarted","Data":"48459f49420cf39e53e718b702999bb29b7d4b0b4236fc3808fd7ff3b87d2c55"} Oct 01 09:51:41 crc kubenswrapper[4787]: I1001 09:51:41.241780 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-de61-account-create-hxhr8" Oct 01 09:51:41 crc kubenswrapper[4787]: I1001 09:51:41.250203 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:51:41 crc kubenswrapper[4787]: I1001 09:51:41.250272 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:51:41 crc kubenswrapper[4787]: I1001 09:51:41.250321 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" Oct 01 09:51:41 crc kubenswrapper[4787]: I1001 09:51:41.251685 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d423136db78a796507e81b181b34f0022442acd501d56cba0addad4da7455409"} pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 09:51:41 crc kubenswrapper[4787]: I1001 09:51:41.251768 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" containerID="cri-o://d423136db78a796507e81b181b34f0022442acd501d56cba0addad4da7455409" gracePeriod=600 Oct 01 09:51:41 crc kubenswrapper[4787]: I1001 09:51:41.709379 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-de61-account-create-hxhr8"] Oct 01 09:51:42 crc kubenswrapper[4787]: I1001 09:51:42.231679 4787 generic.go:334] "Generic (PLEG): container finished" podID="1d2457dd-a699-48b4-86d0-69505da5345c" containerID="44d649cfd48ca68d099ccc760ec2c638396f532602b5f8d951457214115f5bd7" exitCode=0 Oct 01 09:51:42 crc kubenswrapper[4787]: I1001 09:51:42.231796 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-de61-account-create-hxhr8" event={"ID":"1d2457dd-a699-48b4-86d0-69505da5345c","Type":"ContainerDied","Data":"44d649cfd48ca68d099ccc760ec2c638396f532602b5f8d951457214115f5bd7"} Oct 01 09:51:42 crc kubenswrapper[4787]: I1001 09:51:42.232239 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-de61-account-create-hxhr8" event={"ID":"1d2457dd-a699-48b4-86d0-69505da5345c","Type":"ContainerStarted","Data":"ff0d0ff63ffc30fd153f0640a6966ba319730067d55b679e2039f5c84faebd50"} Oct 01 09:51:42 crc kubenswrapper[4787]: I1001 09:51:42.237805 4787 generic.go:334] "Generic (PLEG): container finished" podID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerID="d423136db78a796507e81b181b34f0022442acd501d56cba0addad4da7455409" exitCode=0 Oct 01 09:51:42 crc kubenswrapper[4787]: I1001 09:51:42.237874 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" event={"ID":"fa51b883-5f3e-4141-9d4c-37704aac7718","Type":"ContainerDied","Data":"d423136db78a796507e81b181b34f0022442acd501d56cba0addad4da7455409"} Oct 01 09:51:42 crc kubenswrapper[4787]: I1001 09:51:42.237923 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" event={"ID":"fa51b883-5f3e-4141-9d4c-37704aac7718","Type":"ContainerStarted","Data":"86d3a42fa94bb1529366685be8d4c187b77d63f0ed73a16bbe3812ceaa99b7ae"} Oct 01 09:51:42 crc kubenswrapper[4787]: I1001 09:51:42.237949 4787 scope.go:117] "RemoveContainer" containerID="c5956c6b3dfb9dae0b884bab2812d62e85ed0b6d8154a894d187e5889824b51f" Oct 01 09:51:43 crc kubenswrapper[4787]: I1001 09:51:43.222922 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 01 09:51:43 crc kubenswrapper[4787]: I1001 09:51:43.257322 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3432bc16-99b4-431a-aeb1-600e826bbc3e","Type":"ContainerStarted","Data":"8aec77ffca73c874f2a3570681bffbf1705a2986a735163f54d1b3bc7aaa5ecc"} Oct 01 09:51:43 crc kubenswrapper[4787]: I1001 09:51:43.257366 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3432bc16-99b4-431a-aeb1-600e826bbc3e","Type":"ContainerStarted","Data":"841675f812025d8868022f9630c1eb5786b3e4812549427b17abaaa7868d655b"} Oct 01 09:51:43 crc kubenswrapper[4787]: I1001 09:51:43.257377 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3432bc16-99b4-431a-aeb1-600e826bbc3e","Type":"ContainerStarted","Data":"375dbefc1af4a23029c59ba2c65d99a2dbf9f017fd8086ae755c5c1a5725ac89"} Oct 01 09:51:43 crc kubenswrapper[4787]: I1001 09:51:43.571628 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-de61-account-create-hxhr8" Oct 01 09:51:43 crc kubenswrapper[4787]: I1001 09:51:43.682356 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8b2t\" (UniqueName: \"kubernetes.io/projected/1d2457dd-a699-48b4-86d0-69505da5345c-kube-api-access-g8b2t\") pod \"1d2457dd-a699-48b4-86d0-69505da5345c\" (UID: \"1d2457dd-a699-48b4-86d0-69505da5345c\") " Oct 01 09:51:43 crc kubenswrapper[4787]: I1001 09:51:43.688453 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d2457dd-a699-48b4-86d0-69505da5345c-kube-api-access-g8b2t" (OuterVolumeSpecName: "kube-api-access-g8b2t") pod "1d2457dd-a699-48b4-86d0-69505da5345c" (UID: "1d2457dd-a699-48b4-86d0-69505da5345c"). InnerVolumeSpecName "kube-api-access-g8b2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:51:43 crc kubenswrapper[4787]: I1001 09:51:43.784317 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8b2t\" (UniqueName: \"kubernetes.io/projected/1d2457dd-a699-48b4-86d0-69505da5345c-kube-api-access-g8b2t\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:44 crc kubenswrapper[4787]: I1001 09:51:44.267677 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-de61-account-create-hxhr8" event={"ID":"1d2457dd-a699-48b4-86d0-69505da5345c","Type":"ContainerDied","Data":"ff0d0ff63ffc30fd153f0640a6966ba319730067d55b679e2039f5c84faebd50"} Oct 01 09:51:44 crc kubenswrapper[4787]: I1001 09:51:44.267975 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff0d0ff63ffc30fd153f0640a6966ba319730067d55b679e2039f5c84faebd50" Oct 01 09:51:44 crc kubenswrapper[4787]: I1001 09:51:44.267692 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-de61-account-create-hxhr8" Oct 01 09:51:44 crc kubenswrapper[4787]: I1001 09:51:44.269824 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3432bc16-99b4-431a-aeb1-600e826bbc3e","Type":"ContainerStarted","Data":"6be910b096aac46cb4d535ebe3e8d299ca0b0dacdd7fed58263855baa7f1e3a3"} Oct 01 09:51:47 crc kubenswrapper[4787]: I1001 09:51:47.805952 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-l8fqd" podUID="36eb99a8-609a-4af2-9fa7-d0051806659e" containerName="ovn-controller" probeResult="failure" output=< Oct 01 09:51:47 crc kubenswrapper[4787]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 01 09:51:47 crc kubenswrapper[4787]: > Oct 01 09:51:47 crc kubenswrapper[4787]: I1001 09:51:47.830276 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-6htcd" Oct 01 09:51:47 crc kubenswrapper[4787]: I1001 09:51:47.891274 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.176590 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-rblh2"] Oct 01 09:51:48 crc kubenswrapper[4787]: E1001 09:51:48.178317 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d2457dd-a699-48b4-86d0-69505da5345c" containerName="mariadb-account-create" Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.178347 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d2457dd-a699-48b4-86d0-69505da5345c" containerName="mariadb-account-create" Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.178579 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d2457dd-a699-48b4-86d0-69505da5345c" containerName="mariadb-account-create" Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.179261 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-rblh2" Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.185227 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-rblh2"] Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.268733 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.274152 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-8d4s5"] Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.286941 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-8d4s5"] Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.289687 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-8d4s5" Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.365290 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j42gv\" (UniqueName: \"kubernetes.io/projected/9d39dce3-8e56-4d16-ad94-8f582723de4e-kube-api-access-j42gv\") pod \"cinder-db-create-rblh2\" (UID: \"9d39dce3-8e56-4d16-ad94-8f582723de4e\") " pod="openstack/cinder-db-create-rblh2" Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.466643 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j42gv\" (UniqueName: \"kubernetes.io/projected/9d39dce3-8e56-4d16-ad94-8f582723de4e-kube-api-access-j42gv\") pod \"cinder-db-create-rblh2\" (UID: \"9d39dce3-8e56-4d16-ad94-8f582723de4e\") " pod="openstack/cinder-db-create-rblh2" Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.466737 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vtwm\" (UniqueName: \"kubernetes.io/projected/8a294c79-3cc4-4070-86c7-baad1226a08e-kube-api-access-6vtwm\") pod \"barbican-db-create-8d4s5\" (UID: \"8a294c79-3cc4-4070-86c7-baad1226a08e\") " pod="openstack/barbican-db-create-8d4s5" Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.493442 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j42gv\" (UniqueName: \"kubernetes.io/projected/9d39dce3-8e56-4d16-ad94-8f582723de4e-kube-api-access-j42gv\") pod \"cinder-db-create-rblh2\" (UID: \"9d39dce3-8e56-4d16-ad94-8f582723de4e\") " pod="openstack/cinder-db-create-rblh2" Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.528237 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-rblh2" Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.561740 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-jbw9g"] Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.562907 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-jbw9g" Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.568779 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vtwm\" (UniqueName: \"kubernetes.io/projected/8a294c79-3cc4-4070-86c7-baad1226a08e-kube-api-access-6vtwm\") pod \"barbican-db-create-8d4s5\" (UID: \"8a294c79-3cc4-4070-86c7-baad1226a08e\") " pod="openstack/barbican-db-create-8d4s5" Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.570195 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-jbw9g"] Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.599833 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vtwm\" (UniqueName: \"kubernetes.io/projected/8a294c79-3cc4-4070-86c7-baad1226a08e-kube-api-access-6vtwm\") pod \"barbican-db-create-8d4s5\" (UID: \"8a294c79-3cc4-4070-86c7-baad1226a08e\") " pod="openstack/barbican-db-create-8d4s5" Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.615672 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-8d4s5" Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.633266 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-pd2hv"] Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.634330 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-pd2hv" Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.636778 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.636946 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-qlb5k" Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.637204 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.637296 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.657794 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-pd2hv"] Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.670712 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qz6nl\" (UniqueName: \"kubernetes.io/projected/8de5a45d-bdad-4e9b-9385-13ac70b41009-kube-api-access-qz6nl\") pod \"neutron-db-create-jbw9g\" (UID: \"8de5a45d-bdad-4e9b-9385-13ac70b41009\") " pod="openstack/neutron-db-create-jbw9g" Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.772278 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5fd5149-1773-49d8-8612-fe5eb43ddda8-config-data\") pod \"keystone-db-sync-pd2hv\" (UID: \"c5fd5149-1773-49d8-8612-fe5eb43ddda8\") " pod="openstack/keystone-db-sync-pd2hv" Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.772337 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qz6nl\" (UniqueName: \"kubernetes.io/projected/8de5a45d-bdad-4e9b-9385-13ac70b41009-kube-api-access-qz6nl\") pod \"neutron-db-create-jbw9g\" (UID: \"8de5a45d-bdad-4e9b-9385-13ac70b41009\") " pod="openstack/neutron-db-create-jbw9g" Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.772498 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pd26c\" (UniqueName: \"kubernetes.io/projected/c5fd5149-1773-49d8-8612-fe5eb43ddda8-kube-api-access-pd26c\") pod \"keystone-db-sync-pd2hv\" (UID: \"c5fd5149-1773-49d8-8612-fe5eb43ddda8\") " pod="openstack/keystone-db-sync-pd2hv" Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.772609 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5fd5149-1773-49d8-8612-fe5eb43ddda8-combined-ca-bundle\") pod \"keystone-db-sync-pd2hv\" (UID: \"c5fd5149-1773-49d8-8612-fe5eb43ddda8\") " pod="openstack/keystone-db-sync-pd2hv" Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.791680 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qz6nl\" (UniqueName: \"kubernetes.io/projected/8de5a45d-bdad-4e9b-9385-13ac70b41009-kube-api-access-qz6nl\") pod \"neutron-db-create-jbw9g\" (UID: \"8de5a45d-bdad-4e9b-9385-13ac70b41009\") " pod="openstack/neutron-db-create-jbw9g" Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.874105 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5fd5149-1773-49d8-8612-fe5eb43ddda8-config-data\") pod \"keystone-db-sync-pd2hv\" (UID: \"c5fd5149-1773-49d8-8612-fe5eb43ddda8\") " pod="openstack/keystone-db-sync-pd2hv" Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.874226 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pd26c\" (UniqueName: \"kubernetes.io/projected/c5fd5149-1773-49d8-8612-fe5eb43ddda8-kube-api-access-pd26c\") pod \"keystone-db-sync-pd2hv\" (UID: \"c5fd5149-1773-49d8-8612-fe5eb43ddda8\") " pod="openstack/keystone-db-sync-pd2hv" Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.874301 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5fd5149-1773-49d8-8612-fe5eb43ddda8-combined-ca-bundle\") pod \"keystone-db-sync-pd2hv\" (UID: \"c5fd5149-1773-49d8-8612-fe5eb43ddda8\") " pod="openstack/keystone-db-sync-pd2hv" Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.878425 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5fd5149-1773-49d8-8612-fe5eb43ddda8-combined-ca-bundle\") pod \"keystone-db-sync-pd2hv\" (UID: \"c5fd5149-1773-49d8-8612-fe5eb43ddda8\") " pod="openstack/keystone-db-sync-pd2hv" Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.878708 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5fd5149-1773-49d8-8612-fe5eb43ddda8-config-data\") pod \"keystone-db-sync-pd2hv\" (UID: \"c5fd5149-1773-49d8-8612-fe5eb43ddda8\") " pod="openstack/keystone-db-sync-pd2hv" Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.885470 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-jbw9g" Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.900711 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pd26c\" (UniqueName: \"kubernetes.io/projected/c5fd5149-1773-49d8-8612-fe5eb43ddda8-kube-api-access-pd26c\") pod \"keystone-db-sync-pd2hv\" (UID: \"c5fd5149-1773-49d8-8612-fe5eb43ddda8\") " pod="openstack/keystone-db-sync-pd2hv" Oct 01 09:51:48 crc kubenswrapper[4787]: I1001 09:51:48.959510 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-pd2hv" Oct 01 09:51:50 crc kubenswrapper[4787]: I1001 09:51:50.675856 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-8d4s5"] Oct 01 09:51:50 crc kubenswrapper[4787]: I1001 09:51:50.697002 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-pd2hv"] Oct 01 09:51:50 crc kubenswrapper[4787]: I1001 09:51:50.799227 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-jbw9g"] Oct 01 09:51:50 crc kubenswrapper[4787]: W1001 09:51:50.800931 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8de5a45d_bdad_4e9b_9385_13ac70b41009.slice/crio-017fa4a298f42dcb58212effdd22a8f91bb9b97cb295c0ce7766f11c455fb3fa WatchSource:0}: Error finding container 017fa4a298f42dcb58212effdd22a8f91bb9b97cb295c0ce7766f11c455fb3fa: Status 404 returned error can't find the container with id 017fa4a298f42dcb58212effdd22a8f91bb9b97cb295c0ce7766f11c455fb3fa Oct 01 09:51:50 crc kubenswrapper[4787]: I1001 09:51:50.889945 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-rblh2"] Oct 01 09:51:51 crc kubenswrapper[4787]: I1001 09:51:51.332902 4787 generic.go:334] "Generic (PLEG): container finished" podID="8a294c79-3cc4-4070-86c7-baad1226a08e" containerID="13664b5c7d47d027839c3d62ffea9e5f2cada72768847edbf9afb9e57d3c6c80" exitCode=0 Oct 01 09:51:51 crc kubenswrapper[4787]: I1001 09:51:51.332978 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-8d4s5" event={"ID":"8a294c79-3cc4-4070-86c7-baad1226a08e","Type":"ContainerDied","Data":"13664b5c7d47d027839c3d62ffea9e5f2cada72768847edbf9afb9e57d3c6c80"} Oct 01 09:51:51 crc kubenswrapper[4787]: I1001 09:51:51.333041 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-8d4s5" event={"ID":"8a294c79-3cc4-4070-86c7-baad1226a08e","Type":"ContainerStarted","Data":"49baba670cecfd051e5d66062dfdf215a2cc02ef401804c6d947ba27d26bcbfa"} Oct 01 09:51:51 crc kubenswrapper[4787]: I1001 09:51:51.334517 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-kckrj" event={"ID":"f7251286-7de7-48ef-a418-6d67749e40b8","Type":"ContainerStarted","Data":"ffbc1e8fe63ccb959ba0dd0a30701a2b032acb532b2d1bdb5ffa014ef6d9b226"} Oct 01 09:51:51 crc kubenswrapper[4787]: I1001 09:51:51.336536 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-pd2hv" event={"ID":"c5fd5149-1773-49d8-8612-fe5eb43ddda8","Type":"ContainerStarted","Data":"1d14edf6e9e57c86f9dd52c8cf998206d66d4456465c480ac89adbce98f05806"} Oct 01 09:51:51 crc kubenswrapper[4787]: I1001 09:51:51.337855 4787 generic.go:334] "Generic (PLEG): container finished" podID="8de5a45d-bdad-4e9b-9385-13ac70b41009" containerID="54dc15565137169bed13359225ad074892e4991a14fa5bee94ec079f11e584e9" exitCode=0 Oct 01 09:51:51 crc kubenswrapper[4787]: I1001 09:51:51.337901 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-jbw9g" event={"ID":"8de5a45d-bdad-4e9b-9385-13ac70b41009","Type":"ContainerDied","Data":"54dc15565137169bed13359225ad074892e4991a14fa5bee94ec079f11e584e9"} Oct 01 09:51:51 crc kubenswrapper[4787]: I1001 09:51:51.337921 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-jbw9g" event={"ID":"8de5a45d-bdad-4e9b-9385-13ac70b41009","Type":"ContainerStarted","Data":"017fa4a298f42dcb58212effdd22a8f91bb9b97cb295c0ce7766f11c455fb3fa"} Oct 01 09:51:51 crc kubenswrapper[4787]: I1001 09:51:51.339544 4787 generic.go:334] "Generic (PLEG): container finished" podID="9d39dce3-8e56-4d16-ad94-8f582723de4e" containerID="615c6ef799483af41a0c23e5042cd9bad9e09072c68ddd6ea5acb03343c9db67" exitCode=0 Oct 01 09:51:51 crc kubenswrapper[4787]: I1001 09:51:51.339579 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-rblh2" event={"ID":"9d39dce3-8e56-4d16-ad94-8f582723de4e","Type":"ContainerDied","Data":"615c6ef799483af41a0c23e5042cd9bad9e09072c68ddd6ea5acb03343c9db67"} Oct 01 09:51:51 crc kubenswrapper[4787]: I1001 09:51:51.339598 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-rblh2" event={"ID":"9d39dce3-8e56-4d16-ad94-8f582723de4e","Type":"ContainerStarted","Data":"6a273f41c9ffc488e326a132f74579ca71a8ac1beed59f8df8c81125c6f8536e"} Oct 01 09:51:51 crc kubenswrapper[4787]: I1001 09:51:51.369281 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-kckrj" podStartSLOduration=2.804086287 podStartE2EDuration="15.369260572s" podCreationTimestamp="2025-10-01 09:51:36 +0000 UTC" firstStartedPulling="2025-10-01 09:51:37.684315342 +0000 UTC m=+929.799459499" lastFinishedPulling="2025-10-01 09:51:50.249489627 +0000 UTC m=+942.364633784" observedRunningTime="2025-10-01 09:51:51.363115125 +0000 UTC m=+943.478259302" watchObservedRunningTime="2025-10-01 09:51:51.369260572 +0000 UTC m=+943.484404729" Oct 01 09:51:52 crc kubenswrapper[4787]: I1001 09:51:52.808897 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-l8fqd" podUID="36eb99a8-609a-4af2-9fa7-d0051806659e" containerName="ovn-controller" probeResult="failure" output=< Oct 01 09:51:52 crc kubenswrapper[4787]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 01 09:51:52 crc kubenswrapper[4787]: > Oct 01 09:51:52 crc kubenswrapper[4787]: I1001 09:51:52.840896 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-6htcd" Oct 01 09:51:53 crc kubenswrapper[4787]: I1001 09:51:53.075848 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-l8fqd-config-jhc7t"] Oct 01 09:51:53 crc kubenswrapper[4787]: I1001 09:51:53.082871 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-l8fqd-config-jhc7t" Oct 01 09:51:53 crc kubenswrapper[4787]: I1001 09:51:53.084923 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-l8fqd-config-jhc7t"] Oct 01 09:51:53 crc kubenswrapper[4787]: I1001 09:51:53.087892 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 01 09:51:53 crc kubenswrapper[4787]: I1001 09:51:53.255683 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/daa636bc-1d38-414c-9f72-db5e5fcf0d90-var-run\") pod \"ovn-controller-l8fqd-config-jhc7t\" (UID: \"daa636bc-1d38-414c-9f72-db5e5fcf0d90\") " pod="openstack/ovn-controller-l8fqd-config-jhc7t" Oct 01 09:51:53 crc kubenswrapper[4787]: I1001 09:51:53.255744 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/daa636bc-1d38-414c-9f72-db5e5fcf0d90-additional-scripts\") pod \"ovn-controller-l8fqd-config-jhc7t\" (UID: \"daa636bc-1d38-414c-9f72-db5e5fcf0d90\") " pod="openstack/ovn-controller-l8fqd-config-jhc7t" Oct 01 09:51:53 crc kubenswrapper[4787]: I1001 09:51:53.255771 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/daa636bc-1d38-414c-9f72-db5e5fcf0d90-var-run-ovn\") pod \"ovn-controller-l8fqd-config-jhc7t\" (UID: \"daa636bc-1d38-414c-9f72-db5e5fcf0d90\") " pod="openstack/ovn-controller-l8fqd-config-jhc7t" Oct 01 09:51:53 crc kubenswrapper[4787]: I1001 09:51:53.255945 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/daa636bc-1d38-414c-9f72-db5e5fcf0d90-scripts\") pod \"ovn-controller-l8fqd-config-jhc7t\" (UID: \"daa636bc-1d38-414c-9f72-db5e5fcf0d90\") " pod="openstack/ovn-controller-l8fqd-config-jhc7t" Oct 01 09:51:53 crc kubenswrapper[4787]: I1001 09:51:53.256220 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/daa636bc-1d38-414c-9f72-db5e5fcf0d90-var-log-ovn\") pod \"ovn-controller-l8fqd-config-jhc7t\" (UID: \"daa636bc-1d38-414c-9f72-db5e5fcf0d90\") " pod="openstack/ovn-controller-l8fqd-config-jhc7t" Oct 01 09:51:53 crc kubenswrapper[4787]: I1001 09:51:53.256300 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48x5q\" (UniqueName: \"kubernetes.io/projected/daa636bc-1d38-414c-9f72-db5e5fcf0d90-kube-api-access-48x5q\") pod \"ovn-controller-l8fqd-config-jhc7t\" (UID: \"daa636bc-1d38-414c-9f72-db5e5fcf0d90\") " pod="openstack/ovn-controller-l8fqd-config-jhc7t" Oct 01 09:51:53 crc kubenswrapper[4787]: I1001 09:51:53.357933 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/daa636bc-1d38-414c-9f72-db5e5fcf0d90-additional-scripts\") pod \"ovn-controller-l8fqd-config-jhc7t\" (UID: \"daa636bc-1d38-414c-9f72-db5e5fcf0d90\") " pod="openstack/ovn-controller-l8fqd-config-jhc7t" Oct 01 09:51:53 crc kubenswrapper[4787]: I1001 09:51:53.358015 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/daa636bc-1d38-414c-9f72-db5e5fcf0d90-var-run-ovn\") pod \"ovn-controller-l8fqd-config-jhc7t\" (UID: \"daa636bc-1d38-414c-9f72-db5e5fcf0d90\") " pod="openstack/ovn-controller-l8fqd-config-jhc7t" Oct 01 09:51:53 crc kubenswrapper[4787]: I1001 09:51:53.358072 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/daa636bc-1d38-414c-9f72-db5e5fcf0d90-scripts\") pod \"ovn-controller-l8fqd-config-jhc7t\" (UID: \"daa636bc-1d38-414c-9f72-db5e5fcf0d90\") " pod="openstack/ovn-controller-l8fqd-config-jhc7t" Oct 01 09:51:53 crc kubenswrapper[4787]: I1001 09:51:53.358144 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/daa636bc-1d38-414c-9f72-db5e5fcf0d90-var-log-ovn\") pod \"ovn-controller-l8fqd-config-jhc7t\" (UID: \"daa636bc-1d38-414c-9f72-db5e5fcf0d90\") " pod="openstack/ovn-controller-l8fqd-config-jhc7t" Oct 01 09:51:53 crc kubenswrapper[4787]: I1001 09:51:53.358172 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48x5q\" (UniqueName: \"kubernetes.io/projected/daa636bc-1d38-414c-9f72-db5e5fcf0d90-kube-api-access-48x5q\") pod \"ovn-controller-l8fqd-config-jhc7t\" (UID: \"daa636bc-1d38-414c-9f72-db5e5fcf0d90\") " pod="openstack/ovn-controller-l8fqd-config-jhc7t" Oct 01 09:51:53 crc kubenswrapper[4787]: I1001 09:51:53.358245 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/daa636bc-1d38-414c-9f72-db5e5fcf0d90-var-run\") pod \"ovn-controller-l8fqd-config-jhc7t\" (UID: \"daa636bc-1d38-414c-9f72-db5e5fcf0d90\") " pod="openstack/ovn-controller-l8fqd-config-jhc7t" Oct 01 09:51:53 crc kubenswrapper[4787]: I1001 09:51:53.358551 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/daa636bc-1d38-414c-9f72-db5e5fcf0d90-var-log-ovn\") pod \"ovn-controller-l8fqd-config-jhc7t\" (UID: \"daa636bc-1d38-414c-9f72-db5e5fcf0d90\") " pod="openstack/ovn-controller-l8fqd-config-jhc7t" Oct 01 09:51:53 crc kubenswrapper[4787]: I1001 09:51:53.358580 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/daa636bc-1d38-414c-9f72-db5e5fcf0d90-var-run-ovn\") pod \"ovn-controller-l8fqd-config-jhc7t\" (UID: \"daa636bc-1d38-414c-9f72-db5e5fcf0d90\") " pod="openstack/ovn-controller-l8fqd-config-jhc7t" Oct 01 09:51:53 crc kubenswrapper[4787]: I1001 09:51:53.358564 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/daa636bc-1d38-414c-9f72-db5e5fcf0d90-var-run\") pod \"ovn-controller-l8fqd-config-jhc7t\" (UID: \"daa636bc-1d38-414c-9f72-db5e5fcf0d90\") " pod="openstack/ovn-controller-l8fqd-config-jhc7t" Oct 01 09:51:53 crc kubenswrapper[4787]: I1001 09:51:53.359057 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/daa636bc-1d38-414c-9f72-db5e5fcf0d90-additional-scripts\") pod \"ovn-controller-l8fqd-config-jhc7t\" (UID: \"daa636bc-1d38-414c-9f72-db5e5fcf0d90\") " pod="openstack/ovn-controller-l8fqd-config-jhc7t" Oct 01 09:51:53 crc kubenswrapper[4787]: I1001 09:51:53.360270 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/daa636bc-1d38-414c-9f72-db5e5fcf0d90-scripts\") pod \"ovn-controller-l8fqd-config-jhc7t\" (UID: \"daa636bc-1d38-414c-9f72-db5e5fcf0d90\") " pod="openstack/ovn-controller-l8fqd-config-jhc7t" Oct 01 09:51:53 crc kubenswrapper[4787]: I1001 09:51:53.393188 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48x5q\" (UniqueName: \"kubernetes.io/projected/daa636bc-1d38-414c-9f72-db5e5fcf0d90-kube-api-access-48x5q\") pod \"ovn-controller-l8fqd-config-jhc7t\" (UID: \"daa636bc-1d38-414c-9f72-db5e5fcf0d90\") " pod="openstack/ovn-controller-l8fqd-config-jhc7t" Oct 01 09:51:53 crc kubenswrapper[4787]: I1001 09:51:53.399618 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-l8fqd-config-jhc7t" Oct 01 09:51:53 crc kubenswrapper[4787]: E1001 09:51:53.623449 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.io/podified-antelope-centos9/openstack-swift-container@sha256:7118cc3a695fead2a8bab14c8ace018ed7a5ba23ef347bf4ead44219e8467866: reading manifest sha256:7118cc3a695fead2a8bab14c8ace018ed7a5ba23ef347bf4ead44219e8467866 in quay.io/podified-antelope-centos9/openstack-swift-container: received unexpected HTTP status: 504 Gateway Time-out" image="quay.io/podified-antelope-centos9/openstack-swift-container@sha256:7118cc3a695fead2a8bab14c8ace018ed7a5ba23ef347bf4ead44219e8467866" Oct 01 09:51:53 crc kubenswrapper[4787]: E1001 09:51:53.623605 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:container-server,Image:quay.io/podified-antelope-centos9/openstack-swift-container@sha256:7118cc3a695fead2a8bab14c8ace018ed7a5ba23ef347bf4ead44219e8467866,Command:[/usr/bin/swift-container-server /etc/swift/container-server.conf.d -v],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:container,HostPort:0,ContainerPort:6201,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5b7h56h9dh94h67bh697h95h55hbh555h556h675h5fdh57dh579h5fbh64fh5c9h687hb6h678h5d4h549h54h98h8ch564h5bh5bch55dhc8hf8q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:swift,ReadOnly:false,MountPath:/srv/node/pv,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:etc-swift,ReadOnly:false,MountPath:/etc/swift,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cache,ReadOnly:false,MountPath:/var/cache/swift,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:lock,ReadOnly:false,MountPath:/var/lock,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vzdqq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42445,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-storage-0_openstack(3432bc16-99b4-431a-aeb1-600e826bbc3e): ErrImagePull: initializing source docker://quay.io/podified-antelope-centos9/openstack-swift-container@sha256:7118cc3a695fead2a8bab14c8ace018ed7a5ba23ef347bf4ead44219e8467866: reading manifest sha256:7118cc3a695fead2a8bab14c8ace018ed7a5ba23ef347bf4ead44219e8467866 in quay.io/podified-antelope-centos9/openstack-swift-container: received unexpected HTTP status: 504 Gateway Time-out" logger="UnhandledError" Oct 01 09:51:55 crc kubenswrapper[4787]: I1001 09:51:55.315563 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-rblh2" Oct 01 09:51:55 crc kubenswrapper[4787]: I1001 09:51:55.326421 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-8d4s5" Oct 01 09:51:55 crc kubenswrapper[4787]: I1001 09:51:55.353801 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-jbw9g" Oct 01 09:51:55 crc kubenswrapper[4787]: I1001 09:51:55.386555 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-jbw9g" event={"ID":"8de5a45d-bdad-4e9b-9385-13ac70b41009","Type":"ContainerDied","Data":"017fa4a298f42dcb58212effdd22a8f91bb9b97cb295c0ce7766f11c455fb3fa"} Oct 01 09:51:55 crc kubenswrapper[4787]: I1001 09:51:55.386593 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="017fa4a298f42dcb58212effdd22a8f91bb9b97cb295c0ce7766f11c455fb3fa" Oct 01 09:51:55 crc kubenswrapper[4787]: I1001 09:51:55.386593 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-jbw9g" Oct 01 09:51:55 crc kubenswrapper[4787]: I1001 09:51:55.389838 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-rblh2" Oct 01 09:51:55 crc kubenswrapper[4787]: I1001 09:51:55.389840 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-rblh2" event={"ID":"9d39dce3-8e56-4d16-ad94-8f582723de4e","Type":"ContainerDied","Data":"6a273f41c9ffc488e326a132f74579ca71a8ac1beed59f8df8c81125c6f8536e"} Oct 01 09:51:55 crc kubenswrapper[4787]: I1001 09:51:55.389882 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a273f41c9ffc488e326a132f74579ca71a8ac1beed59f8df8c81125c6f8536e" Oct 01 09:51:55 crc kubenswrapper[4787]: I1001 09:51:55.391419 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-8d4s5" event={"ID":"8a294c79-3cc4-4070-86c7-baad1226a08e","Type":"ContainerDied","Data":"49baba670cecfd051e5d66062dfdf215a2cc02ef401804c6d947ba27d26bcbfa"} Oct 01 09:51:55 crc kubenswrapper[4787]: I1001 09:51:55.391444 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-8d4s5" Oct 01 09:51:55 crc kubenswrapper[4787]: I1001 09:51:55.391450 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="49baba670cecfd051e5d66062dfdf215a2cc02ef401804c6d947ba27d26bcbfa" Oct 01 09:51:55 crc kubenswrapper[4787]: I1001 09:51:55.423553 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j42gv\" (UniqueName: \"kubernetes.io/projected/9d39dce3-8e56-4d16-ad94-8f582723de4e-kube-api-access-j42gv\") pod \"9d39dce3-8e56-4d16-ad94-8f582723de4e\" (UID: \"9d39dce3-8e56-4d16-ad94-8f582723de4e\") " Oct 01 09:51:55 crc kubenswrapper[4787]: I1001 09:51:55.423765 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qz6nl\" (UniqueName: \"kubernetes.io/projected/8de5a45d-bdad-4e9b-9385-13ac70b41009-kube-api-access-qz6nl\") pod \"8de5a45d-bdad-4e9b-9385-13ac70b41009\" (UID: \"8de5a45d-bdad-4e9b-9385-13ac70b41009\") " Oct 01 09:51:55 crc kubenswrapper[4787]: I1001 09:51:55.423806 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vtwm\" (UniqueName: \"kubernetes.io/projected/8a294c79-3cc4-4070-86c7-baad1226a08e-kube-api-access-6vtwm\") pod \"8a294c79-3cc4-4070-86c7-baad1226a08e\" (UID: \"8a294c79-3cc4-4070-86c7-baad1226a08e\") " Oct 01 09:51:55 crc kubenswrapper[4787]: I1001 09:51:55.428988 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a294c79-3cc4-4070-86c7-baad1226a08e-kube-api-access-6vtwm" (OuterVolumeSpecName: "kube-api-access-6vtwm") pod "8a294c79-3cc4-4070-86c7-baad1226a08e" (UID: "8a294c79-3cc4-4070-86c7-baad1226a08e"). InnerVolumeSpecName "kube-api-access-6vtwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:51:55 crc kubenswrapper[4787]: I1001 09:51:55.429374 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8de5a45d-bdad-4e9b-9385-13ac70b41009-kube-api-access-qz6nl" (OuterVolumeSpecName: "kube-api-access-qz6nl") pod "8de5a45d-bdad-4e9b-9385-13ac70b41009" (UID: "8de5a45d-bdad-4e9b-9385-13ac70b41009"). InnerVolumeSpecName "kube-api-access-qz6nl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:51:55 crc kubenswrapper[4787]: I1001 09:51:55.429652 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d39dce3-8e56-4d16-ad94-8f582723de4e-kube-api-access-j42gv" (OuterVolumeSpecName: "kube-api-access-j42gv") pod "9d39dce3-8e56-4d16-ad94-8f582723de4e" (UID: "9d39dce3-8e56-4d16-ad94-8f582723de4e"). InnerVolumeSpecName "kube-api-access-j42gv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:51:55 crc kubenswrapper[4787]: I1001 09:51:55.525395 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qz6nl\" (UniqueName: \"kubernetes.io/projected/8de5a45d-bdad-4e9b-9385-13ac70b41009-kube-api-access-qz6nl\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:55 crc kubenswrapper[4787]: I1001 09:51:55.525432 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6vtwm\" (UniqueName: \"kubernetes.io/projected/8a294c79-3cc4-4070-86c7-baad1226a08e-kube-api-access-6vtwm\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:55 crc kubenswrapper[4787]: I1001 09:51:55.525445 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j42gv\" (UniqueName: \"kubernetes.io/projected/9d39dce3-8e56-4d16-ad94-8f582723de4e-kube-api-access-j42gv\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:55 crc kubenswrapper[4787]: I1001 09:51:55.562304 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-l8fqd-config-jhc7t"] Oct 01 09:51:55 crc kubenswrapper[4787]: W1001 09:51:55.575900 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddaa636bc_1d38_414c_9f72_db5e5fcf0d90.slice/crio-2b14d4eb5e8f0d1d75aa49c2aa2b242b7fe23f719c5a0bc72cf7dd10a3493521 WatchSource:0}: Error finding container 2b14d4eb5e8f0d1d75aa49c2aa2b242b7fe23f719c5a0bc72cf7dd10a3493521: Status 404 returned error can't find the container with id 2b14d4eb5e8f0d1d75aa49c2aa2b242b7fe23f719c5a0bc72cf7dd10a3493521 Oct 01 09:51:56 crc kubenswrapper[4787]: I1001 09:51:56.404501 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-pd2hv" event={"ID":"c5fd5149-1773-49d8-8612-fe5eb43ddda8","Type":"ContainerStarted","Data":"36087873618e96f830b48167fbcbb872b6a1cae4fbd26be0edf8d3a2923a3560"} Oct 01 09:51:56 crc kubenswrapper[4787]: I1001 09:51:56.409393 4787 generic.go:334] "Generic (PLEG): container finished" podID="daa636bc-1d38-414c-9f72-db5e5fcf0d90" containerID="df38295b98c61b89830a5e9c42a88dd5e5dbafc3ddb80edc86f22e38179d976b" exitCode=0 Oct 01 09:51:56 crc kubenswrapper[4787]: I1001 09:51:56.409451 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-l8fqd-config-jhc7t" event={"ID":"daa636bc-1d38-414c-9f72-db5e5fcf0d90","Type":"ContainerDied","Data":"df38295b98c61b89830a5e9c42a88dd5e5dbafc3ddb80edc86f22e38179d976b"} Oct 01 09:51:56 crc kubenswrapper[4787]: I1001 09:51:56.409519 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-l8fqd-config-jhc7t" event={"ID":"daa636bc-1d38-414c-9f72-db5e5fcf0d90","Type":"ContainerStarted","Data":"2b14d4eb5e8f0d1d75aa49c2aa2b242b7fe23f719c5a0bc72cf7dd10a3493521"} Oct 01 09:51:56 crc kubenswrapper[4787]: I1001 09:51:56.422524 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-pd2hv" podStartSLOduration=3.999924413 podStartE2EDuration="8.42250166s" podCreationTimestamp="2025-10-01 09:51:48 +0000 UTC" firstStartedPulling="2025-10-01 09:51:50.730556424 +0000 UTC m=+942.845700581" lastFinishedPulling="2025-10-01 09:51:55.153133671 +0000 UTC m=+947.268277828" observedRunningTime="2025-10-01 09:51:56.420387176 +0000 UTC m=+948.535531343" watchObservedRunningTime="2025-10-01 09:51:56.42250166 +0000 UTC m=+948.537645817" Oct 01 09:51:57 crc kubenswrapper[4787]: I1001 09:51:57.420007 4787 generic.go:334] "Generic (PLEG): container finished" podID="f7251286-7de7-48ef-a418-6d67749e40b8" containerID="ffbc1e8fe63ccb959ba0dd0a30701a2b032acb532b2d1bdb5ffa014ef6d9b226" exitCode=0 Oct 01 09:51:57 crc kubenswrapper[4787]: I1001 09:51:57.420122 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-kckrj" event={"ID":"f7251286-7de7-48ef-a418-6d67749e40b8","Type":"ContainerDied","Data":"ffbc1e8fe63ccb959ba0dd0a30701a2b032acb532b2d1bdb5ffa014ef6d9b226"} Oct 01 09:51:57 crc kubenswrapper[4787]: I1001 09:51:57.775627 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-l8fqd-config-jhc7t" Oct 01 09:51:57 crc kubenswrapper[4787]: I1001 09:51:57.825125 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-l8fqd" Oct 01 09:51:57 crc kubenswrapper[4787]: I1001 09:51:57.876575 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/daa636bc-1d38-414c-9f72-db5e5fcf0d90-var-run" (OuterVolumeSpecName: "var-run") pod "daa636bc-1d38-414c-9f72-db5e5fcf0d90" (UID: "daa636bc-1d38-414c-9f72-db5e5fcf0d90"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:51:57 crc kubenswrapper[4787]: I1001 09:51:57.877684 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/daa636bc-1d38-414c-9f72-db5e5fcf0d90-var-run\") pod \"daa636bc-1d38-414c-9f72-db5e5fcf0d90\" (UID: \"daa636bc-1d38-414c-9f72-db5e5fcf0d90\") " Oct 01 09:51:57 crc kubenswrapper[4787]: I1001 09:51:57.877767 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/daa636bc-1d38-414c-9f72-db5e5fcf0d90-scripts\") pod \"daa636bc-1d38-414c-9f72-db5e5fcf0d90\" (UID: \"daa636bc-1d38-414c-9f72-db5e5fcf0d90\") " Oct 01 09:51:57 crc kubenswrapper[4787]: I1001 09:51:57.877798 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-48x5q\" (UniqueName: \"kubernetes.io/projected/daa636bc-1d38-414c-9f72-db5e5fcf0d90-kube-api-access-48x5q\") pod \"daa636bc-1d38-414c-9f72-db5e5fcf0d90\" (UID: \"daa636bc-1d38-414c-9f72-db5e5fcf0d90\") " Oct 01 09:51:57 crc kubenswrapper[4787]: I1001 09:51:57.877831 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/daa636bc-1d38-414c-9f72-db5e5fcf0d90-additional-scripts\") pod \"daa636bc-1d38-414c-9f72-db5e5fcf0d90\" (UID: \"daa636bc-1d38-414c-9f72-db5e5fcf0d90\") " Oct 01 09:51:57 crc kubenswrapper[4787]: I1001 09:51:57.877891 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/daa636bc-1d38-414c-9f72-db5e5fcf0d90-var-log-ovn\") pod \"daa636bc-1d38-414c-9f72-db5e5fcf0d90\" (UID: \"daa636bc-1d38-414c-9f72-db5e5fcf0d90\") " Oct 01 09:51:57 crc kubenswrapper[4787]: I1001 09:51:57.877944 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/daa636bc-1d38-414c-9f72-db5e5fcf0d90-var-run-ovn\") pod \"daa636bc-1d38-414c-9f72-db5e5fcf0d90\" (UID: \"daa636bc-1d38-414c-9f72-db5e5fcf0d90\") " Oct 01 09:51:57 crc kubenswrapper[4787]: I1001 09:51:57.878557 4787 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/daa636bc-1d38-414c-9f72-db5e5fcf0d90-var-run\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:57 crc kubenswrapper[4787]: I1001 09:51:57.879156 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/daa636bc-1d38-414c-9f72-db5e5fcf0d90-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "daa636bc-1d38-414c-9f72-db5e5fcf0d90" (UID: "daa636bc-1d38-414c-9f72-db5e5fcf0d90"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:51:57 crc kubenswrapper[4787]: I1001 09:51:57.879838 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/daa636bc-1d38-414c-9f72-db5e5fcf0d90-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "daa636bc-1d38-414c-9f72-db5e5fcf0d90" (UID: "daa636bc-1d38-414c-9f72-db5e5fcf0d90"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:51:57 crc kubenswrapper[4787]: I1001 09:51:57.880377 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/daa636bc-1d38-414c-9f72-db5e5fcf0d90-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "daa636bc-1d38-414c-9f72-db5e5fcf0d90" (UID: "daa636bc-1d38-414c-9f72-db5e5fcf0d90"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:51:57 crc kubenswrapper[4787]: I1001 09:51:57.881674 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/daa636bc-1d38-414c-9f72-db5e5fcf0d90-scripts" (OuterVolumeSpecName: "scripts") pod "daa636bc-1d38-414c-9f72-db5e5fcf0d90" (UID: "daa636bc-1d38-414c-9f72-db5e5fcf0d90"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:51:57 crc kubenswrapper[4787]: I1001 09:51:57.885859 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/daa636bc-1d38-414c-9f72-db5e5fcf0d90-kube-api-access-48x5q" (OuterVolumeSpecName: "kube-api-access-48x5q") pod "daa636bc-1d38-414c-9f72-db5e5fcf0d90" (UID: "daa636bc-1d38-414c-9f72-db5e5fcf0d90"). InnerVolumeSpecName "kube-api-access-48x5q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:51:57 crc kubenswrapper[4787]: I1001 09:51:57.980026 4787 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/daa636bc-1d38-414c-9f72-db5e5fcf0d90-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:57 crc kubenswrapper[4787]: I1001 09:51:57.980431 4787 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/daa636bc-1d38-414c-9f72-db5e5fcf0d90-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:57 crc kubenswrapper[4787]: I1001 09:51:57.980524 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/daa636bc-1d38-414c-9f72-db5e5fcf0d90-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:57 crc kubenswrapper[4787]: I1001 09:51:57.980591 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-48x5q\" (UniqueName: \"kubernetes.io/projected/daa636bc-1d38-414c-9f72-db5e5fcf0d90-kube-api-access-48x5q\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:57 crc kubenswrapper[4787]: I1001 09:51:57.980665 4787 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/daa636bc-1d38-414c-9f72-db5e5fcf0d90-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:58 crc kubenswrapper[4787]: I1001 09:51:58.431555 4787 generic.go:334] "Generic (PLEG): container finished" podID="c5fd5149-1773-49d8-8612-fe5eb43ddda8" containerID="36087873618e96f830b48167fbcbb872b6a1cae4fbd26be0edf8d3a2923a3560" exitCode=0 Oct 01 09:51:58 crc kubenswrapper[4787]: I1001 09:51:58.431661 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-pd2hv" event={"ID":"c5fd5149-1773-49d8-8612-fe5eb43ddda8","Type":"ContainerDied","Data":"36087873618e96f830b48167fbcbb872b6a1cae4fbd26be0edf8d3a2923a3560"} Oct 01 09:51:58 crc kubenswrapper[4787]: I1001 09:51:58.434107 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-l8fqd-config-jhc7t" event={"ID":"daa636bc-1d38-414c-9f72-db5e5fcf0d90","Type":"ContainerDied","Data":"2b14d4eb5e8f0d1d75aa49c2aa2b242b7fe23f719c5a0bc72cf7dd10a3493521"} Oct 01 09:51:58 crc kubenswrapper[4787]: I1001 09:51:58.434259 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b14d4eb5e8f0d1d75aa49c2aa2b242b7fe23f719c5a0bc72cf7dd10a3493521" Oct 01 09:51:58 crc kubenswrapper[4787]: I1001 09:51:58.434149 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-l8fqd-config-jhc7t" Oct 01 09:51:58 crc kubenswrapper[4787]: I1001 09:51:58.840986 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-kckrj" Oct 01 09:51:58 crc kubenswrapper[4787]: I1001 09:51:58.877733 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-l8fqd-config-jhc7t"] Oct 01 09:51:58 crc kubenswrapper[4787]: I1001 09:51:58.887528 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-l8fqd-config-jhc7t"] Oct 01 09:51:58 crc kubenswrapper[4787]: I1001 09:51:58.894837 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2ff5\" (UniqueName: \"kubernetes.io/projected/f7251286-7de7-48ef-a418-6d67749e40b8-kube-api-access-d2ff5\") pod \"f7251286-7de7-48ef-a418-6d67749e40b8\" (UID: \"f7251286-7de7-48ef-a418-6d67749e40b8\") " Oct 01 09:51:58 crc kubenswrapper[4787]: I1001 09:51:58.894927 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f7251286-7de7-48ef-a418-6d67749e40b8-db-sync-config-data\") pod \"f7251286-7de7-48ef-a418-6d67749e40b8\" (UID: \"f7251286-7de7-48ef-a418-6d67749e40b8\") " Oct 01 09:51:58 crc kubenswrapper[4787]: I1001 09:51:58.894961 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7251286-7de7-48ef-a418-6d67749e40b8-config-data\") pod \"f7251286-7de7-48ef-a418-6d67749e40b8\" (UID: \"f7251286-7de7-48ef-a418-6d67749e40b8\") " Oct 01 09:51:58 crc kubenswrapper[4787]: I1001 09:51:58.895034 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7251286-7de7-48ef-a418-6d67749e40b8-combined-ca-bundle\") pod \"f7251286-7de7-48ef-a418-6d67749e40b8\" (UID: \"f7251286-7de7-48ef-a418-6d67749e40b8\") " Oct 01 09:51:58 crc kubenswrapper[4787]: I1001 09:51:58.899477 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7251286-7de7-48ef-a418-6d67749e40b8-kube-api-access-d2ff5" (OuterVolumeSpecName: "kube-api-access-d2ff5") pod "f7251286-7de7-48ef-a418-6d67749e40b8" (UID: "f7251286-7de7-48ef-a418-6d67749e40b8"). InnerVolumeSpecName "kube-api-access-d2ff5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:51:58 crc kubenswrapper[4787]: I1001 09:51:58.906518 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7251286-7de7-48ef-a418-6d67749e40b8-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "f7251286-7de7-48ef-a418-6d67749e40b8" (UID: "f7251286-7de7-48ef-a418-6d67749e40b8"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:51:58 crc kubenswrapper[4787]: I1001 09:51:58.943217 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7251286-7de7-48ef-a418-6d67749e40b8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f7251286-7de7-48ef-a418-6d67749e40b8" (UID: "f7251286-7de7-48ef-a418-6d67749e40b8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:51:58 crc kubenswrapper[4787]: I1001 09:51:58.945704 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-l8fqd-config-jw2wk"] Oct 01 09:51:58 crc kubenswrapper[4787]: E1001 09:51:58.946039 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="daa636bc-1d38-414c-9f72-db5e5fcf0d90" containerName="ovn-config" Oct 01 09:51:58 crc kubenswrapper[4787]: I1001 09:51:58.946059 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="daa636bc-1d38-414c-9f72-db5e5fcf0d90" containerName="ovn-config" Oct 01 09:51:58 crc kubenswrapper[4787]: E1001 09:51:58.946093 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d39dce3-8e56-4d16-ad94-8f582723de4e" containerName="mariadb-database-create" Oct 01 09:51:58 crc kubenswrapper[4787]: I1001 09:51:58.946101 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d39dce3-8e56-4d16-ad94-8f582723de4e" containerName="mariadb-database-create" Oct 01 09:51:58 crc kubenswrapper[4787]: E1001 09:51:58.946118 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a294c79-3cc4-4070-86c7-baad1226a08e" containerName="mariadb-database-create" Oct 01 09:51:58 crc kubenswrapper[4787]: I1001 09:51:58.946132 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a294c79-3cc4-4070-86c7-baad1226a08e" containerName="mariadb-database-create" Oct 01 09:51:58 crc kubenswrapper[4787]: E1001 09:51:58.946153 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7251286-7de7-48ef-a418-6d67749e40b8" containerName="glance-db-sync" Oct 01 09:51:58 crc kubenswrapper[4787]: I1001 09:51:58.946160 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7251286-7de7-48ef-a418-6d67749e40b8" containerName="glance-db-sync" Oct 01 09:51:58 crc kubenswrapper[4787]: E1001 09:51:58.946176 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8de5a45d-bdad-4e9b-9385-13ac70b41009" containerName="mariadb-database-create" Oct 01 09:51:58 crc kubenswrapper[4787]: I1001 09:51:58.946183 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="8de5a45d-bdad-4e9b-9385-13ac70b41009" containerName="mariadb-database-create" Oct 01 09:51:58 crc kubenswrapper[4787]: I1001 09:51:58.946385 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7251286-7de7-48ef-a418-6d67749e40b8" containerName="glance-db-sync" Oct 01 09:51:58 crc kubenswrapper[4787]: I1001 09:51:58.946413 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a294c79-3cc4-4070-86c7-baad1226a08e" containerName="mariadb-database-create" Oct 01 09:51:58 crc kubenswrapper[4787]: I1001 09:51:58.946433 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="8de5a45d-bdad-4e9b-9385-13ac70b41009" containerName="mariadb-database-create" Oct 01 09:51:58 crc kubenswrapper[4787]: I1001 09:51:58.946451 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="daa636bc-1d38-414c-9f72-db5e5fcf0d90" containerName="ovn-config" Oct 01 09:51:58 crc kubenswrapper[4787]: I1001 09:51:58.946462 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d39dce3-8e56-4d16-ad94-8f582723de4e" containerName="mariadb-database-create" Oct 01 09:51:58 crc kubenswrapper[4787]: I1001 09:51:58.947166 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-l8fqd-config-jw2wk" Oct 01 09:51:58 crc kubenswrapper[4787]: I1001 09:51:58.949854 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 01 09:51:58 crc kubenswrapper[4787]: I1001 09:51:58.968342 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7251286-7de7-48ef-a418-6d67749e40b8-config-data" (OuterVolumeSpecName: "config-data") pod "f7251286-7de7-48ef-a418-6d67749e40b8" (UID: "f7251286-7de7-48ef-a418-6d67749e40b8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:51:58 crc kubenswrapper[4787]: I1001 09:51:58.982567 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-l8fqd-config-jw2wk"] Oct 01 09:51:58 crc kubenswrapper[4787]: I1001 09:51:58.998132 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/80bd8ecf-a2fb-4635-83f1-050e227e74e6-var-run-ovn\") pod \"ovn-controller-l8fqd-config-jw2wk\" (UID: \"80bd8ecf-a2fb-4635-83f1-050e227e74e6\") " pod="openstack/ovn-controller-l8fqd-config-jw2wk" Oct 01 09:51:58 crc kubenswrapper[4787]: I1001 09:51:58.998187 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26wbl\" (UniqueName: \"kubernetes.io/projected/80bd8ecf-a2fb-4635-83f1-050e227e74e6-kube-api-access-26wbl\") pod \"ovn-controller-l8fqd-config-jw2wk\" (UID: \"80bd8ecf-a2fb-4635-83f1-050e227e74e6\") " pod="openstack/ovn-controller-l8fqd-config-jw2wk" Oct 01 09:51:58 crc kubenswrapper[4787]: I1001 09:51:58.998229 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/80bd8ecf-a2fb-4635-83f1-050e227e74e6-var-run\") pod \"ovn-controller-l8fqd-config-jw2wk\" (UID: \"80bd8ecf-a2fb-4635-83f1-050e227e74e6\") " pod="openstack/ovn-controller-l8fqd-config-jw2wk" Oct 01 09:51:58 crc kubenswrapper[4787]: I1001 09:51:58.998325 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/80bd8ecf-a2fb-4635-83f1-050e227e74e6-var-log-ovn\") pod \"ovn-controller-l8fqd-config-jw2wk\" (UID: \"80bd8ecf-a2fb-4635-83f1-050e227e74e6\") " pod="openstack/ovn-controller-l8fqd-config-jw2wk" Oct 01 09:51:58 crc kubenswrapper[4787]: I1001 09:51:58.998360 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/80bd8ecf-a2fb-4635-83f1-050e227e74e6-additional-scripts\") pod \"ovn-controller-l8fqd-config-jw2wk\" (UID: \"80bd8ecf-a2fb-4635-83f1-050e227e74e6\") " pod="openstack/ovn-controller-l8fqd-config-jw2wk" Oct 01 09:51:58 crc kubenswrapper[4787]: I1001 09:51:58.998384 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/80bd8ecf-a2fb-4635-83f1-050e227e74e6-scripts\") pod \"ovn-controller-l8fqd-config-jw2wk\" (UID: \"80bd8ecf-a2fb-4635-83f1-050e227e74e6\") " pod="openstack/ovn-controller-l8fqd-config-jw2wk" Oct 01 09:51:58 crc kubenswrapper[4787]: I1001 09:51:58.998431 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2ff5\" (UniqueName: \"kubernetes.io/projected/f7251286-7de7-48ef-a418-6d67749e40b8-kube-api-access-d2ff5\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:58 crc kubenswrapper[4787]: I1001 09:51:58.998443 4787 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f7251286-7de7-48ef-a418-6d67749e40b8-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:58 crc kubenswrapper[4787]: I1001 09:51:58.998452 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7251286-7de7-48ef-a418-6d67749e40b8-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:58 crc kubenswrapper[4787]: I1001 09:51:58.998461 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7251286-7de7-48ef-a418-6d67749e40b8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:59 crc kubenswrapper[4787]: I1001 09:51:59.100740 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/80bd8ecf-a2fb-4635-83f1-050e227e74e6-var-run\") pod \"ovn-controller-l8fqd-config-jw2wk\" (UID: \"80bd8ecf-a2fb-4635-83f1-050e227e74e6\") " pod="openstack/ovn-controller-l8fqd-config-jw2wk" Oct 01 09:51:59 crc kubenswrapper[4787]: I1001 09:51:59.101016 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/80bd8ecf-a2fb-4635-83f1-050e227e74e6-var-log-ovn\") pod \"ovn-controller-l8fqd-config-jw2wk\" (UID: \"80bd8ecf-a2fb-4635-83f1-050e227e74e6\") " pod="openstack/ovn-controller-l8fqd-config-jw2wk" Oct 01 09:51:59 crc kubenswrapper[4787]: I1001 09:51:59.101114 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/80bd8ecf-a2fb-4635-83f1-050e227e74e6-additional-scripts\") pod \"ovn-controller-l8fqd-config-jw2wk\" (UID: \"80bd8ecf-a2fb-4635-83f1-050e227e74e6\") " pod="openstack/ovn-controller-l8fqd-config-jw2wk" Oct 01 09:51:59 crc kubenswrapper[4787]: I1001 09:51:59.101167 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/80bd8ecf-a2fb-4635-83f1-050e227e74e6-scripts\") pod \"ovn-controller-l8fqd-config-jw2wk\" (UID: \"80bd8ecf-a2fb-4635-83f1-050e227e74e6\") " pod="openstack/ovn-controller-l8fqd-config-jw2wk" Oct 01 09:51:59 crc kubenswrapper[4787]: I1001 09:51:59.101222 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/80bd8ecf-a2fb-4635-83f1-050e227e74e6-var-run-ovn\") pod \"ovn-controller-l8fqd-config-jw2wk\" (UID: \"80bd8ecf-a2fb-4635-83f1-050e227e74e6\") " pod="openstack/ovn-controller-l8fqd-config-jw2wk" Oct 01 09:51:59 crc kubenswrapper[4787]: I1001 09:51:59.101243 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/80bd8ecf-a2fb-4635-83f1-050e227e74e6-var-run\") pod \"ovn-controller-l8fqd-config-jw2wk\" (UID: \"80bd8ecf-a2fb-4635-83f1-050e227e74e6\") " pod="openstack/ovn-controller-l8fqd-config-jw2wk" Oct 01 09:51:59 crc kubenswrapper[4787]: I1001 09:51:59.101276 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26wbl\" (UniqueName: \"kubernetes.io/projected/80bd8ecf-a2fb-4635-83f1-050e227e74e6-kube-api-access-26wbl\") pod \"ovn-controller-l8fqd-config-jw2wk\" (UID: \"80bd8ecf-a2fb-4635-83f1-050e227e74e6\") " pod="openstack/ovn-controller-l8fqd-config-jw2wk" Oct 01 09:51:59 crc kubenswrapper[4787]: I1001 09:51:59.101331 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/80bd8ecf-a2fb-4635-83f1-050e227e74e6-var-log-ovn\") pod \"ovn-controller-l8fqd-config-jw2wk\" (UID: \"80bd8ecf-a2fb-4635-83f1-050e227e74e6\") " pod="openstack/ovn-controller-l8fqd-config-jw2wk" Oct 01 09:51:59 crc kubenswrapper[4787]: I1001 09:51:59.101522 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/80bd8ecf-a2fb-4635-83f1-050e227e74e6-var-run-ovn\") pod \"ovn-controller-l8fqd-config-jw2wk\" (UID: \"80bd8ecf-a2fb-4635-83f1-050e227e74e6\") " pod="openstack/ovn-controller-l8fqd-config-jw2wk" Oct 01 09:51:59 crc kubenswrapper[4787]: I1001 09:51:59.102149 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/80bd8ecf-a2fb-4635-83f1-050e227e74e6-additional-scripts\") pod \"ovn-controller-l8fqd-config-jw2wk\" (UID: \"80bd8ecf-a2fb-4635-83f1-050e227e74e6\") " pod="openstack/ovn-controller-l8fqd-config-jw2wk" Oct 01 09:51:59 crc kubenswrapper[4787]: I1001 09:51:59.103800 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/80bd8ecf-a2fb-4635-83f1-050e227e74e6-scripts\") pod \"ovn-controller-l8fqd-config-jw2wk\" (UID: \"80bd8ecf-a2fb-4635-83f1-050e227e74e6\") " pod="openstack/ovn-controller-l8fqd-config-jw2wk" Oct 01 09:51:59 crc kubenswrapper[4787]: I1001 09:51:59.122350 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26wbl\" (UniqueName: \"kubernetes.io/projected/80bd8ecf-a2fb-4635-83f1-050e227e74e6-kube-api-access-26wbl\") pod \"ovn-controller-l8fqd-config-jw2wk\" (UID: \"80bd8ecf-a2fb-4635-83f1-050e227e74e6\") " pod="openstack/ovn-controller-l8fqd-config-jw2wk" Oct 01 09:51:59 crc kubenswrapper[4787]: I1001 09:51:59.325201 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-l8fqd-config-jw2wk" Oct 01 09:51:59 crc kubenswrapper[4787]: I1001 09:51:59.458427 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-kckrj" Oct 01 09:51:59 crc kubenswrapper[4787]: I1001 09:51:59.459575 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-kckrj" event={"ID":"f7251286-7de7-48ef-a418-6d67749e40b8","Type":"ContainerDied","Data":"09a402eafdfd770a947714471641ca787f0259aa361fa442756a49dfbd67ef7f"} Oct 01 09:51:59 crc kubenswrapper[4787]: I1001 09:51:59.459626 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="09a402eafdfd770a947714471641ca787f0259aa361fa442756a49dfbd67ef7f" Oct 01 09:51:59 crc kubenswrapper[4787]: I1001 09:51:59.749141 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-pd2hv" Oct 01 09:51:59 crc kubenswrapper[4787]: I1001 09:51:59.812422 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5fd5149-1773-49d8-8612-fe5eb43ddda8-combined-ca-bundle\") pod \"c5fd5149-1773-49d8-8612-fe5eb43ddda8\" (UID: \"c5fd5149-1773-49d8-8612-fe5eb43ddda8\") " Oct 01 09:51:59 crc kubenswrapper[4787]: I1001 09:51:59.812511 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5fd5149-1773-49d8-8612-fe5eb43ddda8-config-data\") pod \"c5fd5149-1773-49d8-8612-fe5eb43ddda8\" (UID: \"c5fd5149-1773-49d8-8612-fe5eb43ddda8\") " Oct 01 09:51:59 crc kubenswrapper[4787]: I1001 09:51:59.812532 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pd26c\" (UniqueName: \"kubernetes.io/projected/c5fd5149-1773-49d8-8612-fe5eb43ddda8-kube-api-access-pd26c\") pod \"c5fd5149-1773-49d8-8612-fe5eb43ddda8\" (UID: \"c5fd5149-1773-49d8-8612-fe5eb43ddda8\") " Oct 01 09:51:59 crc kubenswrapper[4787]: I1001 09:51:59.817767 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-757b4b6d4f-5l2dw"] Oct 01 09:51:59 crc kubenswrapper[4787]: E1001 09:51:59.818190 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5fd5149-1773-49d8-8612-fe5eb43ddda8" containerName="keystone-db-sync" Oct 01 09:51:59 crc kubenswrapper[4787]: I1001 09:51:59.818208 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5fd5149-1773-49d8-8612-fe5eb43ddda8" containerName="keystone-db-sync" Oct 01 09:51:59 crc kubenswrapper[4787]: I1001 09:51:59.818427 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5fd5149-1773-49d8-8612-fe5eb43ddda8" containerName="keystone-db-sync" Oct 01 09:51:59 crc kubenswrapper[4787]: I1001 09:51:59.831323 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5fd5149-1773-49d8-8612-fe5eb43ddda8-kube-api-access-pd26c" (OuterVolumeSpecName: "kube-api-access-pd26c") pod "c5fd5149-1773-49d8-8612-fe5eb43ddda8" (UID: "c5fd5149-1773-49d8-8612-fe5eb43ddda8"). InnerVolumeSpecName "kube-api-access-pd26c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:51:59 crc kubenswrapper[4787]: I1001 09:51:59.844168 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4b6d4f-5l2dw" Oct 01 09:51:59 crc kubenswrapper[4787]: I1001 09:51:59.869884 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4b6d4f-5l2dw"] Oct 01 09:51:59 crc kubenswrapper[4787]: I1001 09:51:59.915638 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6754353d-92dc-4f0b-8e19-4ac10e7210d7-dns-svc\") pod \"dnsmasq-dns-757b4b6d4f-5l2dw\" (UID: \"6754353d-92dc-4f0b-8e19-4ac10e7210d7\") " pod="openstack/dnsmasq-dns-757b4b6d4f-5l2dw" Oct 01 09:51:59 crc kubenswrapper[4787]: I1001 09:51:59.915789 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xz2fw\" (UniqueName: \"kubernetes.io/projected/6754353d-92dc-4f0b-8e19-4ac10e7210d7-kube-api-access-xz2fw\") pod \"dnsmasq-dns-757b4b6d4f-5l2dw\" (UID: \"6754353d-92dc-4f0b-8e19-4ac10e7210d7\") " pod="openstack/dnsmasq-dns-757b4b6d4f-5l2dw" Oct 01 09:51:59 crc kubenswrapper[4787]: I1001 09:51:59.915861 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6754353d-92dc-4f0b-8e19-4ac10e7210d7-config\") pod \"dnsmasq-dns-757b4b6d4f-5l2dw\" (UID: \"6754353d-92dc-4f0b-8e19-4ac10e7210d7\") " pod="openstack/dnsmasq-dns-757b4b6d4f-5l2dw" Oct 01 09:51:59 crc kubenswrapper[4787]: I1001 09:51:59.915880 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6754353d-92dc-4f0b-8e19-4ac10e7210d7-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4b6d4f-5l2dw\" (UID: \"6754353d-92dc-4f0b-8e19-4ac10e7210d7\") " pod="openstack/dnsmasq-dns-757b4b6d4f-5l2dw" Oct 01 09:51:59 crc kubenswrapper[4787]: I1001 09:51:59.915941 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6754353d-92dc-4f0b-8e19-4ac10e7210d7-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4b6d4f-5l2dw\" (UID: \"6754353d-92dc-4f0b-8e19-4ac10e7210d7\") " pod="openstack/dnsmasq-dns-757b4b6d4f-5l2dw" Oct 01 09:51:59 crc kubenswrapper[4787]: I1001 09:51:59.916019 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pd26c\" (UniqueName: \"kubernetes.io/projected/c5fd5149-1773-49d8-8612-fe5eb43ddda8-kube-api-access-pd26c\") on node \"crc\" DevicePath \"\"" Oct 01 09:51:59 crc kubenswrapper[4787]: I1001 09:51:59.925213 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-l8fqd-config-jw2wk"] Oct 01 09:51:59 crc kubenswrapper[4787]: I1001 09:51:59.969479 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5fd5149-1773-49d8-8612-fe5eb43ddda8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c5fd5149-1773-49d8-8612-fe5eb43ddda8" (UID: "c5fd5149-1773-49d8-8612-fe5eb43ddda8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:51:59 crc kubenswrapper[4787]: I1001 09:51:59.984334 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5fd5149-1773-49d8-8612-fe5eb43ddda8-config-data" (OuterVolumeSpecName: "config-data") pod "c5fd5149-1773-49d8-8612-fe5eb43ddda8" (UID: "c5fd5149-1773-49d8-8612-fe5eb43ddda8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.020964 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6754353d-92dc-4f0b-8e19-4ac10e7210d7-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4b6d4f-5l2dw\" (UID: \"6754353d-92dc-4f0b-8e19-4ac10e7210d7\") " pod="openstack/dnsmasq-dns-757b4b6d4f-5l2dw" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.021040 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6754353d-92dc-4f0b-8e19-4ac10e7210d7-dns-svc\") pod \"dnsmasq-dns-757b4b6d4f-5l2dw\" (UID: \"6754353d-92dc-4f0b-8e19-4ac10e7210d7\") " pod="openstack/dnsmasq-dns-757b4b6d4f-5l2dw" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.021114 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xz2fw\" (UniqueName: \"kubernetes.io/projected/6754353d-92dc-4f0b-8e19-4ac10e7210d7-kube-api-access-xz2fw\") pod \"dnsmasq-dns-757b4b6d4f-5l2dw\" (UID: \"6754353d-92dc-4f0b-8e19-4ac10e7210d7\") " pod="openstack/dnsmasq-dns-757b4b6d4f-5l2dw" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.021165 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6754353d-92dc-4f0b-8e19-4ac10e7210d7-config\") pod \"dnsmasq-dns-757b4b6d4f-5l2dw\" (UID: \"6754353d-92dc-4f0b-8e19-4ac10e7210d7\") " pod="openstack/dnsmasq-dns-757b4b6d4f-5l2dw" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.021180 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6754353d-92dc-4f0b-8e19-4ac10e7210d7-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4b6d4f-5l2dw\" (UID: \"6754353d-92dc-4f0b-8e19-4ac10e7210d7\") " pod="openstack/dnsmasq-dns-757b4b6d4f-5l2dw" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.021231 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5fd5149-1773-49d8-8612-fe5eb43ddda8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.021241 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5fd5149-1773-49d8-8612-fe5eb43ddda8-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.022222 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6754353d-92dc-4f0b-8e19-4ac10e7210d7-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4b6d4f-5l2dw\" (UID: \"6754353d-92dc-4f0b-8e19-4ac10e7210d7\") " pod="openstack/dnsmasq-dns-757b4b6d4f-5l2dw" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.022390 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6754353d-92dc-4f0b-8e19-4ac10e7210d7-dns-svc\") pod \"dnsmasq-dns-757b4b6d4f-5l2dw\" (UID: \"6754353d-92dc-4f0b-8e19-4ac10e7210d7\") " pod="openstack/dnsmasq-dns-757b4b6d4f-5l2dw" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.023066 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6754353d-92dc-4f0b-8e19-4ac10e7210d7-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4b6d4f-5l2dw\" (UID: \"6754353d-92dc-4f0b-8e19-4ac10e7210d7\") " pod="openstack/dnsmasq-dns-757b4b6d4f-5l2dw" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.023113 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6754353d-92dc-4f0b-8e19-4ac10e7210d7-config\") pod \"dnsmasq-dns-757b4b6d4f-5l2dw\" (UID: \"6754353d-92dc-4f0b-8e19-4ac10e7210d7\") " pod="openstack/dnsmasq-dns-757b4b6d4f-5l2dw" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.059627 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xz2fw\" (UniqueName: \"kubernetes.io/projected/6754353d-92dc-4f0b-8e19-4ac10e7210d7-kube-api-access-xz2fw\") pod \"dnsmasq-dns-757b4b6d4f-5l2dw\" (UID: \"6754353d-92dc-4f0b-8e19-4ac10e7210d7\") " pod="openstack/dnsmasq-dns-757b4b6d4f-5l2dw" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.328631 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4b6d4f-5l2dw" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.491990 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-pd2hv" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.492260 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-pd2hv" event={"ID":"c5fd5149-1773-49d8-8612-fe5eb43ddda8","Type":"ContainerDied","Data":"1d14edf6e9e57c86f9dd52c8cf998206d66d4456465c480ac89adbce98f05806"} Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.492333 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d14edf6e9e57c86f9dd52c8cf998206d66d4456465c480ac89adbce98f05806" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.505223 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-l8fqd-config-jw2wk" event={"ID":"80bd8ecf-a2fb-4635-83f1-050e227e74e6","Type":"ContainerStarted","Data":"02b4037afb172a39d39693d2e624c65b1326bc9b665ead63181fab2ff8180c5c"} Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.505274 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-l8fqd-config-jw2wk" event={"ID":"80bd8ecf-a2fb-4635-83f1-050e227e74e6","Type":"ContainerStarted","Data":"5afc08c7207c715e1f38eebf786d7ad575fa03c0ea2fcada177b81cca344542e"} Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.550342 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="daa636bc-1d38-414c-9f72-db5e5fcf0d90" path="/var/lib/kubelet/pods/daa636bc-1d38-414c-9f72-db5e5fcf0d90/volumes" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.557000 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-l8fqd-config-jw2wk" podStartSLOduration=2.556982423 podStartE2EDuration="2.556982423s" podCreationTimestamp="2025-10-01 09:51:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:52:00.52913721 +0000 UTC m=+952.644281377" watchObservedRunningTime="2025-10-01 09:52:00.556982423 +0000 UTC m=+952.672126580" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.707764 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4b6d4f-5l2dw"] Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.741017 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cd86cbb59-nqs8t"] Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.748386 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cd86cbb59-nqs8t" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.785798 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-tk8w8"] Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.787391 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-tk8w8" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.795182 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cd86cbb59-nqs8t"] Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.800398 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.800621 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.800724 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-qlb5k" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.800821 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.807006 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-tk8w8"] Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.844829 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/266f85d9-996e-41a9-a83d-5451e74689b6-scripts\") pod \"keystone-bootstrap-tk8w8\" (UID: \"266f85d9-996e-41a9-a83d-5451e74689b6\") " pod="openstack/keystone-bootstrap-tk8w8" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.844913 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/266f85d9-996e-41a9-a83d-5451e74689b6-config-data\") pod \"keystone-bootstrap-tk8w8\" (UID: \"266f85d9-996e-41a9-a83d-5451e74689b6\") " pod="openstack/keystone-bootstrap-tk8w8" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.844930 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9c3e6671-985a-4fc7-a8a0-17f290f6a382-ovsdbserver-sb\") pod \"dnsmasq-dns-7cd86cbb59-nqs8t\" (UID: \"9c3e6671-985a-4fc7-a8a0-17f290f6a382\") " pod="openstack/dnsmasq-dns-7cd86cbb59-nqs8t" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.844957 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c3e6671-985a-4fc7-a8a0-17f290f6a382-config\") pod \"dnsmasq-dns-7cd86cbb59-nqs8t\" (UID: \"9c3e6671-985a-4fc7-a8a0-17f290f6a382\") " pod="openstack/dnsmasq-dns-7cd86cbb59-nqs8t" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.844977 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/266f85d9-996e-41a9-a83d-5451e74689b6-fernet-keys\") pod \"keystone-bootstrap-tk8w8\" (UID: \"266f85d9-996e-41a9-a83d-5451e74689b6\") " pod="openstack/keystone-bootstrap-tk8w8" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.845005 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9c3e6671-985a-4fc7-a8a0-17f290f6a382-dns-svc\") pod \"dnsmasq-dns-7cd86cbb59-nqs8t\" (UID: \"9c3e6671-985a-4fc7-a8a0-17f290f6a382\") " pod="openstack/dnsmasq-dns-7cd86cbb59-nqs8t" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.845024 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9c3e6671-985a-4fc7-a8a0-17f290f6a382-ovsdbserver-nb\") pod \"dnsmasq-dns-7cd86cbb59-nqs8t\" (UID: \"9c3e6671-985a-4fc7-a8a0-17f290f6a382\") " pod="openstack/dnsmasq-dns-7cd86cbb59-nqs8t" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.845045 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/266f85d9-996e-41a9-a83d-5451e74689b6-combined-ca-bundle\") pod \"keystone-bootstrap-tk8w8\" (UID: \"266f85d9-996e-41a9-a83d-5451e74689b6\") " pod="openstack/keystone-bootstrap-tk8w8" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.845066 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/266f85d9-996e-41a9-a83d-5451e74689b6-credential-keys\") pod \"keystone-bootstrap-tk8w8\" (UID: \"266f85d9-996e-41a9-a83d-5451e74689b6\") " pod="openstack/keystone-bootstrap-tk8w8" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.845096 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfkhh\" (UniqueName: \"kubernetes.io/projected/266f85d9-996e-41a9-a83d-5451e74689b6-kube-api-access-jfkhh\") pod \"keystone-bootstrap-tk8w8\" (UID: \"266f85d9-996e-41a9-a83d-5451e74689b6\") " pod="openstack/keystone-bootstrap-tk8w8" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.845121 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nhtc\" (UniqueName: \"kubernetes.io/projected/9c3e6671-985a-4fc7-a8a0-17f290f6a382-kube-api-access-5nhtc\") pod \"dnsmasq-dns-7cd86cbb59-nqs8t\" (UID: \"9c3e6671-985a-4fc7-a8a0-17f290f6a382\") " pod="openstack/dnsmasq-dns-7cd86cbb59-nqs8t" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.854248 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4b6d4f-5l2dw"] Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.946156 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/266f85d9-996e-41a9-a83d-5451e74689b6-config-data\") pod \"keystone-bootstrap-tk8w8\" (UID: \"266f85d9-996e-41a9-a83d-5451e74689b6\") " pod="openstack/keystone-bootstrap-tk8w8" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.946480 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9c3e6671-985a-4fc7-a8a0-17f290f6a382-ovsdbserver-sb\") pod \"dnsmasq-dns-7cd86cbb59-nqs8t\" (UID: \"9c3e6671-985a-4fc7-a8a0-17f290f6a382\") " pod="openstack/dnsmasq-dns-7cd86cbb59-nqs8t" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.946523 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c3e6671-985a-4fc7-a8a0-17f290f6a382-config\") pod \"dnsmasq-dns-7cd86cbb59-nqs8t\" (UID: \"9c3e6671-985a-4fc7-a8a0-17f290f6a382\") " pod="openstack/dnsmasq-dns-7cd86cbb59-nqs8t" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.946551 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/266f85d9-996e-41a9-a83d-5451e74689b6-fernet-keys\") pod \"keystone-bootstrap-tk8w8\" (UID: \"266f85d9-996e-41a9-a83d-5451e74689b6\") " pod="openstack/keystone-bootstrap-tk8w8" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.946594 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9c3e6671-985a-4fc7-a8a0-17f290f6a382-dns-svc\") pod \"dnsmasq-dns-7cd86cbb59-nqs8t\" (UID: \"9c3e6671-985a-4fc7-a8a0-17f290f6a382\") " pod="openstack/dnsmasq-dns-7cd86cbb59-nqs8t" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.946610 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9c3e6671-985a-4fc7-a8a0-17f290f6a382-ovsdbserver-nb\") pod \"dnsmasq-dns-7cd86cbb59-nqs8t\" (UID: \"9c3e6671-985a-4fc7-a8a0-17f290f6a382\") " pod="openstack/dnsmasq-dns-7cd86cbb59-nqs8t" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.946656 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/266f85d9-996e-41a9-a83d-5451e74689b6-combined-ca-bundle\") pod \"keystone-bootstrap-tk8w8\" (UID: \"266f85d9-996e-41a9-a83d-5451e74689b6\") " pod="openstack/keystone-bootstrap-tk8w8" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.946684 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/266f85d9-996e-41a9-a83d-5451e74689b6-credential-keys\") pod \"keystone-bootstrap-tk8w8\" (UID: \"266f85d9-996e-41a9-a83d-5451e74689b6\") " pod="openstack/keystone-bootstrap-tk8w8" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.946700 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfkhh\" (UniqueName: \"kubernetes.io/projected/266f85d9-996e-41a9-a83d-5451e74689b6-kube-api-access-jfkhh\") pod \"keystone-bootstrap-tk8w8\" (UID: \"266f85d9-996e-41a9-a83d-5451e74689b6\") " pod="openstack/keystone-bootstrap-tk8w8" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.946737 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nhtc\" (UniqueName: \"kubernetes.io/projected/9c3e6671-985a-4fc7-a8a0-17f290f6a382-kube-api-access-5nhtc\") pod \"dnsmasq-dns-7cd86cbb59-nqs8t\" (UID: \"9c3e6671-985a-4fc7-a8a0-17f290f6a382\") " pod="openstack/dnsmasq-dns-7cd86cbb59-nqs8t" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.946772 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/266f85d9-996e-41a9-a83d-5451e74689b6-scripts\") pod \"keystone-bootstrap-tk8w8\" (UID: \"266f85d9-996e-41a9-a83d-5451e74689b6\") " pod="openstack/keystone-bootstrap-tk8w8" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.948936 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9c3e6671-985a-4fc7-a8a0-17f290f6a382-dns-svc\") pod \"dnsmasq-dns-7cd86cbb59-nqs8t\" (UID: \"9c3e6671-985a-4fc7-a8a0-17f290f6a382\") " pod="openstack/dnsmasq-dns-7cd86cbb59-nqs8t" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.951720 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c3e6671-985a-4fc7-a8a0-17f290f6a382-config\") pod \"dnsmasq-dns-7cd86cbb59-nqs8t\" (UID: \"9c3e6671-985a-4fc7-a8a0-17f290f6a382\") " pod="openstack/dnsmasq-dns-7cd86cbb59-nqs8t" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.956455 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9c3e6671-985a-4fc7-a8a0-17f290f6a382-ovsdbserver-sb\") pod \"dnsmasq-dns-7cd86cbb59-nqs8t\" (UID: \"9c3e6671-985a-4fc7-a8a0-17f290f6a382\") " pod="openstack/dnsmasq-dns-7cd86cbb59-nqs8t" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.959736 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/266f85d9-996e-41a9-a83d-5451e74689b6-fernet-keys\") pod \"keystone-bootstrap-tk8w8\" (UID: \"266f85d9-996e-41a9-a83d-5451e74689b6\") " pod="openstack/keystone-bootstrap-tk8w8" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.959977 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/266f85d9-996e-41a9-a83d-5451e74689b6-config-data\") pod \"keystone-bootstrap-tk8w8\" (UID: \"266f85d9-996e-41a9-a83d-5451e74689b6\") " pod="openstack/keystone-bootstrap-tk8w8" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.960355 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9c3e6671-985a-4fc7-a8a0-17f290f6a382-ovsdbserver-nb\") pod \"dnsmasq-dns-7cd86cbb59-nqs8t\" (UID: \"9c3e6671-985a-4fc7-a8a0-17f290f6a382\") " pod="openstack/dnsmasq-dns-7cd86cbb59-nqs8t" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.961997 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-58779b6b6f-rqbn7"] Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.965154 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-58779b6b6f-rqbn7" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.967835 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/266f85d9-996e-41a9-a83d-5451e74689b6-scripts\") pod \"keystone-bootstrap-tk8w8\" (UID: \"266f85d9-996e-41a9-a83d-5451e74689b6\") " pod="openstack/keystone-bootstrap-tk8w8" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.970380 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/266f85d9-996e-41a9-a83d-5451e74689b6-combined-ca-bundle\") pod \"keystone-bootstrap-tk8w8\" (UID: \"266f85d9-996e-41a9-a83d-5451e74689b6\") " pod="openstack/keystone-bootstrap-tk8w8" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.972382 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.972847 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.972961 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-bbwwh" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.975275 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.978965 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/266f85d9-996e-41a9-a83d-5451e74689b6-credential-keys\") pod \"keystone-bootstrap-tk8w8\" (UID: \"266f85d9-996e-41a9-a83d-5451e74689b6\") " pod="openstack/keystone-bootstrap-tk8w8" Oct 01 09:52:00 crc kubenswrapper[4787]: I1001 09:52:00.984362 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-58779b6b6f-rqbn7"] Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.009929 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nhtc\" (UniqueName: \"kubernetes.io/projected/9c3e6671-985a-4fc7-a8a0-17f290f6a382-kube-api-access-5nhtc\") pod \"dnsmasq-dns-7cd86cbb59-nqs8t\" (UID: \"9c3e6671-985a-4fc7-a8a0-17f290f6a382\") " pod="openstack/dnsmasq-dns-7cd86cbb59-nqs8t" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.016348 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfkhh\" (UniqueName: \"kubernetes.io/projected/266f85d9-996e-41a9-a83d-5451e74689b6-kube-api-access-jfkhh\") pod \"keystone-bootstrap-tk8w8\" (UID: \"266f85d9-996e-41a9-a83d-5451e74689b6\") " pod="openstack/keystone-bootstrap-tk8w8" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.022725 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-tk8w8" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.048994 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3-config-data\") pod \"horizon-58779b6b6f-rqbn7\" (UID: \"2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3\") " pod="openstack/horizon-58779b6b6f-rqbn7" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.049063 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3-logs\") pod \"horizon-58779b6b6f-rqbn7\" (UID: \"2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3\") " pod="openstack/horizon-58779b6b6f-rqbn7" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.049104 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3-horizon-secret-key\") pod \"horizon-58779b6b6f-rqbn7\" (UID: \"2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3\") " pod="openstack/horizon-58779b6b6f-rqbn7" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.049136 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vncx9\" (UniqueName: \"kubernetes.io/projected/2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3-kube-api-access-vncx9\") pod \"horizon-58779b6b6f-rqbn7\" (UID: \"2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3\") " pod="openstack/horizon-58779b6b6f-rqbn7" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.049170 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3-scripts\") pod \"horizon-58779b6b6f-rqbn7\" (UID: \"2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3\") " pod="openstack/horizon-58779b6b6f-rqbn7" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.094516 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.097707 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.100739 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.112518 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.145279 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.152087 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vncx9\" (UniqueName: \"kubernetes.io/projected/2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3-kube-api-access-vncx9\") pod \"horizon-58779b6b6f-rqbn7\" (UID: \"2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3\") " pod="openstack/horizon-58779b6b6f-rqbn7" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.152146 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3-scripts\") pod \"horizon-58779b6b6f-rqbn7\" (UID: \"2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3\") " pod="openstack/horizon-58779b6b6f-rqbn7" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.152193 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cs95k\" (UniqueName: \"kubernetes.io/projected/9587d64e-33e3-45d2-8ee7-a776fcc60d88-kube-api-access-cs95k\") pod \"ceilometer-0\" (UID: \"9587d64e-33e3-45d2-8ee7-a776fcc60d88\") " pod="openstack/ceilometer-0" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.152230 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9587d64e-33e3-45d2-8ee7-a776fcc60d88-scripts\") pod \"ceilometer-0\" (UID: \"9587d64e-33e3-45d2-8ee7-a776fcc60d88\") " pod="openstack/ceilometer-0" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.152275 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9587d64e-33e3-45d2-8ee7-a776fcc60d88-run-httpd\") pod \"ceilometer-0\" (UID: \"9587d64e-33e3-45d2-8ee7-a776fcc60d88\") " pod="openstack/ceilometer-0" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.152322 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3-config-data\") pod \"horizon-58779b6b6f-rqbn7\" (UID: \"2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3\") " pod="openstack/horizon-58779b6b6f-rqbn7" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.152353 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9587d64e-33e3-45d2-8ee7-a776fcc60d88-config-data\") pod \"ceilometer-0\" (UID: \"9587d64e-33e3-45d2-8ee7-a776fcc60d88\") " pod="openstack/ceilometer-0" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.152382 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3-logs\") pod \"horizon-58779b6b6f-rqbn7\" (UID: \"2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3\") " pod="openstack/horizon-58779b6b6f-rqbn7" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.152415 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3-horizon-secret-key\") pod \"horizon-58779b6b6f-rqbn7\" (UID: \"2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3\") " pod="openstack/horizon-58779b6b6f-rqbn7" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.152433 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9587d64e-33e3-45d2-8ee7-a776fcc60d88-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9587d64e-33e3-45d2-8ee7-a776fcc60d88\") " pod="openstack/ceilometer-0" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.152454 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9587d64e-33e3-45d2-8ee7-a776fcc60d88-log-httpd\") pod \"ceilometer-0\" (UID: \"9587d64e-33e3-45d2-8ee7-a776fcc60d88\") " pod="openstack/ceilometer-0" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.152475 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9587d64e-33e3-45d2-8ee7-a776fcc60d88-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9587d64e-33e3-45d2-8ee7-a776fcc60d88\") " pod="openstack/ceilometer-0" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.153403 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3-scripts\") pod \"horizon-58779b6b6f-rqbn7\" (UID: \"2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3\") " pod="openstack/horizon-58779b6b6f-rqbn7" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.154522 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3-logs\") pod \"horizon-58779b6b6f-rqbn7\" (UID: \"2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3\") " pod="openstack/horizon-58779b6b6f-rqbn7" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.155552 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3-config-data\") pod \"horizon-58779b6b6f-rqbn7\" (UID: \"2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3\") " pod="openstack/horizon-58779b6b6f-rqbn7" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.158427 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3-horizon-secret-key\") pod \"horizon-58779b6b6f-rqbn7\" (UID: \"2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3\") " pod="openstack/horizon-58779b6b6f-rqbn7" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.162889 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cd86cbb59-nqs8t" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.197137 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-b48666669-4sqql"] Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.198707 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-b48666669-4sqql" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.213199 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vncx9\" (UniqueName: \"kubernetes.io/projected/2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3-kube-api-access-vncx9\") pod \"horizon-58779b6b6f-rqbn7\" (UID: \"2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3\") " pod="openstack/horizon-58779b6b6f-rqbn7" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.247200 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-b48666669-4sqql"] Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.255558 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3869928b-1191-433d-87bf-7187615b57e0-scripts\") pod \"horizon-b48666669-4sqql\" (UID: \"3869928b-1191-433d-87bf-7187615b57e0\") " pod="openstack/horizon-b48666669-4sqql" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.255621 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9587d64e-33e3-45d2-8ee7-a776fcc60d88-config-data\") pod \"ceilometer-0\" (UID: \"9587d64e-33e3-45d2-8ee7-a776fcc60d88\") " pod="openstack/ceilometer-0" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.255666 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9587d64e-33e3-45d2-8ee7-a776fcc60d88-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9587d64e-33e3-45d2-8ee7-a776fcc60d88\") " pod="openstack/ceilometer-0" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.255690 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9587d64e-33e3-45d2-8ee7-a776fcc60d88-log-httpd\") pod \"ceilometer-0\" (UID: \"9587d64e-33e3-45d2-8ee7-a776fcc60d88\") " pod="openstack/ceilometer-0" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.255708 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9587d64e-33e3-45d2-8ee7-a776fcc60d88-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9587d64e-33e3-45d2-8ee7-a776fcc60d88\") " pod="openstack/ceilometer-0" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.255829 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvxbh\" (UniqueName: \"kubernetes.io/projected/3869928b-1191-433d-87bf-7187615b57e0-kube-api-access-gvxbh\") pod \"horizon-b48666669-4sqql\" (UID: \"3869928b-1191-433d-87bf-7187615b57e0\") " pod="openstack/horizon-b48666669-4sqql" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.255868 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cs95k\" (UniqueName: \"kubernetes.io/projected/9587d64e-33e3-45d2-8ee7-a776fcc60d88-kube-api-access-cs95k\") pod \"ceilometer-0\" (UID: \"9587d64e-33e3-45d2-8ee7-a776fcc60d88\") " pod="openstack/ceilometer-0" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.255889 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3869928b-1191-433d-87bf-7187615b57e0-config-data\") pod \"horizon-b48666669-4sqql\" (UID: \"3869928b-1191-433d-87bf-7187615b57e0\") " pod="openstack/horizon-b48666669-4sqql" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.255913 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9587d64e-33e3-45d2-8ee7-a776fcc60d88-scripts\") pod \"ceilometer-0\" (UID: \"9587d64e-33e3-45d2-8ee7-a776fcc60d88\") " pod="openstack/ceilometer-0" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.255938 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3869928b-1191-433d-87bf-7187615b57e0-horizon-secret-key\") pod \"horizon-b48666669-4sqql\" (UID: \"3869928b-1191-433d-87bf-7187615b57e0\") " pod="openstack/horizon-b48666669-4sqql" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.255955 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3869928b-1191-433d-87bf-7187615b57e0-logs\") pod \"horizon-b48666669-4sqql\" (UID: \"3869928b-1191-433d-87bf-7187615b57e0\") " pod="openstack/horizon-b48666669-4sqql" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.255978 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9587d64e-33e3-45d2-8ee7-a776fcc60d88-run-httpd\") pod \"ceilometer-0\" (UID: \"9587d64e-33e3-45d2-8ee7-a776fcc60d88\") " pod="openstack/ceilometer-0" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.256491 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-slvbz"] Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.257902 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-slvbz" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.259723 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9587d64e-33e3-45d2-8ee7-a776fcc60d88-run-httpd\") pod \"ceilometer-0\" (UID: \"9587d64e-33e3-45d2-8ee7-a776fcc60d88\") " pod="openstack/ceilometer-0" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.263321 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.263564 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-sfxs7" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.263658 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.265137 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9587d64e-33e3-45d2-8ee7-a776fcc60d88-log-httpd\") pod \"ceilometer-0\" (UID: \"9587d64e-33e3-45d2-8ee7-a776fcc60d88\") " pod="openstack/ceilometer-0" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.273666 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9587d64e-33e3-45d2-8ee7-a776fcc60d88-scripts\") pod \"ceilometer-0\" (UID: \"9587d64e-33e3-45d2-8ee7-a776fcc60d88\") " pod="openstack/ceilometer-0" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.275044 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9587d64e-33e3-45d2-8ee7-a776fcc60d88-config-data\") pod \"ceilometer-0\" (UID: \"9587d64e-33e3-45d2-8ee7-a776fcc60d88\") " pod="openstack/ceilometer-0" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.275422 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9587d64e-33e3-45d2-8ee7-a776fcc60d88-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9587d64e-33e3-45d2-8ee7-a776fcc60d88\") " pod="openstack/ceilometer-0" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.277264 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.279953 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.283853 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-bskck" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.283932 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.284099 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.285822 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-slvbz"] Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.288855 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cs95k\" (UniqueName: \"kubernetes.io/projected/9587d64e-33e3-45d2-8ee7-a776fcc60d88-kube-api-access-cs95k\") pod \"ceilometer-0\" (UID: \"9587d64e-33e3-45d2-8ee7-a776fcc60d88\") " pod="openstack/ceilometer-0" Oct 01 09:52:01 crc kubenswrapper[4787]: I1001 09:52:01.295592 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9587d64e-33e3-45d2-8ee7-a776fcc60d88-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9587d64e-33e3-45d2-8ee7-a776fcc60d88\") " pod="openstack/ceilometer-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.310761 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.319999 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cd86cbb59-nqs8t"] Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.356556 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-58779b6b6f-rqbn7" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.368479 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7bb97fccb5-vblx7"] Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.369853 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bb97fccb5-vblx7" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.370339 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6af63266-547b-4537-9290-338b2c0a2d73-logs\") pod \"placement-db-sync-slvbz\" (UID: \"6af63266-547b-4537-9290-338b2c0a2d73\") " pod="openstack/placement-db-sync-slvbz" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.370384 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3869928b-1191-433d-87bf-7187615b57e0-scripts\") pod \"horizon-b48666669-4sqql\" (UID: \"3869928b-1191-433d-87bf-7187615b57e0\") " pod="openstack/horizon-b48666669-4sqql" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.370407 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3-logs\") pod \"glance-default-external-api-0\" (UID: \"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.370436 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3-scripts\") pod \"glance-default-external-api-0\" (UID: \"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.370455 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.370477 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.370545 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6af63266-547b-4537-9290-338b2c0a2d73-config-data\") pod \"placement-db-sync-slvbz\" (UID: \"6af63266-547b-4537-9290-338b2c0a2d73\") " pod="openstack/placement-db-sync-slvbz" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.370572 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6af63266-547b-4537-9290-338b2c0a2d73-scripts\") pod \"placement-db-sync-slvbz\" (UID: \"6af63266-547b-4537-9290-338b2c0a2d73\") " pod="openstack/placement-db-sync-slvbz" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.370617 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvxbh\" (UniqueName: \"kubernetes.io/projected/3869928b-1191-433d-87bf-7187615b57e0-kube-api-access-gvxbh\") pod \"horizon-b48666669-4sqql\" (UID: \"3869928b-1191-433d-87bf-7187615b57e0\") " pod="openstack/horizon-b48666669-4sqql" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.370642 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9pxp\" (UniqueName: \"kubernetes.io/projected/6af63266-547b-4537-9290-338b2c0a2d73-kube-api-access-m9pxp\") pod \"placement-db-sync-slvbz\" (UID: \"6af63266-547b-4537-9290-338b2c0a2d73\") " pod="openstack/placement-db-sync-slvbz" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.370731 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.370751 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6af63266-547b-4537-9290-338b2c0a2d73-combined-ca-bundle\") pod \"placement-db-sync-slvbz\" (UID: \"6af63266-547b-4537-9290-338b2c0a2d73\") " pod="openstack/placement-db-sync-slvbz" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.370778 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3869928b-1191-433d-87bf-7187615b57e0-config-data\") pod \"horizon-b48666669-4sqql\" (UID: \"3869928b-1191-433d-87bf-7187615b57e0\") " pod="openstack/horizon-b48666669-4sqql" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.370810 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3869928b-1191-433d-87bf-7187615b57e0-horizon-secret-key\") pod \"horizon-b48666669-4sqql\" (UID: \"3869928b-1191-433d-87bf-7187615b57e0\") " pod="openstack/horizon-b48666669-4sqql" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.370833 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3869928b-1191-433d-87bf-7187615b57e0-logs\") pod \"horizon-b48666669-4sqql\" (UID: \"3869928b-1191-433d-87bf-7187615b57e0\") " pod="openstack/horizon-b48666669-4sqql" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.370851 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3-config-data\") pod \"glance-default-external-api-0\" (UID: \"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.370882 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84s2j\" (UniqueName: \"kubernetes.io/projected/1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3-kube-api-access-84s2j\") pod \"glance-default-external-api-0\" (UID: \"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.371635 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3869928b-1191-433d-87bf-7187615b57e0-logs\") pod \"horizon-b48666669-4sqql\" (UID: \"3869928b-1191-433d-87bf-7187615b57e0\") " pod="openstack/horizon-b48666669-4sqql" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.372058 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3869928b-1191-433d-87bf-7187615b57e0-scripts\") pod \"horizon-b48666669-4sqql\" (UID: \"3869928b-1191-433d-87bf-7187615b57e0\") " pod="openstack/horizon-b48666669-4sqql" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.373516 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3869928b-1191-433d-87bf-7187615b57e0-config-data\") pod \"horizon-b48666669-4sqql\" (UID: \"3869928b-1191-433d-87bf-7187615b57e0\") " pod="openstack/horizon-b48666669-4sqql" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.377093 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3869928b-1191-433d-87bf-7187615b57e0-horizon-secret-key\") pod \"horizon-b48666669-4sqql\" (UID: \"3869928b-1191-433d-87bf-7187615b57e0\") " pod="openstack/horizon-b48666669-4sqql" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.379159 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bb97fccb5-vblx7"] Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.397906 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.399500 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.401720 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.402334 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvxbh\" (UniqueName: \"kubernetes.io/projected/3869928b-1191-433d-87bf-7187615b57e0-kube-api-access-gvxbh\") pod \"horizon-b48666669-4sqql\" (UID: \"3869928b-1191-433d-87bf-7187615b57e0\") " pod="openstack/horizon-b48666669-4sqql" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.417172 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.437928 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.473528 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3-logs\") pod \"glance-default-external-api-0\" (UID: \"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.473569 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a63d659d-52a4-4f33-85c2-1461423776a3-dns-svc\") pod \"dnsmasq-dns-7bb97fccb5-vblx7\" (UID: \"a63d659d-52a4-4f33-85c2-1461423776a3\") " pod="openstack/dnsmasq-dns-7bb97fccb5-vblx7" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.473593 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qllv9\" (UniqueName: \"kubernetes.io/projected/55091ec1-5a26-4825-9a61-a49cecda5a6f-kube-api-access-qllv9\") pod \"glance-default-internal-api-0\" (UID: \"55091ec1-5a26-4825-9a61-a49cecda5a6f\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.473611 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a63d659d-52a4-4f33-85c2-1461423776a3-config\") pod \"dnsmasq-dns-7bb97fccb5-vblx7\" (UID: \"a63d659d-52a4-4f33-85c2-1461423776a3\") " pod="openstack/dnsmasq-dns-7bb97fccb5-vblx7" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.473628 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a63d659d-52a4-4f33-85c2-1461423776a3-ovsdbserver-sb\") pod \"dnsmasq-dns-7bb97fccb5-vblx7\" (UID: \"a63d659d-52a4-4f33-85c2-1461423776a3\") " pod="openstack/dnsmasq-dns-7bb97fccb5-vblx7" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.473651 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3-scripts\") pod \"glance-default-external-api-0\" (UID: \"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.473672 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.473693 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.473724 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a63d659d-52a4-4f33-85c2-1461423776a3-ovsdbserver-nb\") pod \"dnsmasq-dns-7bb97fccb5-vblx7\" (UID: \"a63d659d-52a4-4f33-85c2-1461423776a3\") " pod="openstack/dnsmasq-dns-7bb97fccb5-vblx7" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.473743 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/55091ec1-5a26-4825-9a61-a49cecda5a6f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"55091ec1-5a26-4825-9a61-a49cecda5a6f\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.473757 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55091ec1-5a26-4825-9a61-a49cecda5a6f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"55091ec1-5a26-4825-9a61-a49cecda5a6f\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.473776 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"55091ec1-5a26-4825-9a61-a49cecda5a6f\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.473794 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6af63266-547b-4537-9290-338b2c0a2d73-config-data\") pod \"placement-db-sync-slvbz\" (UID: \"6af63266-547b-4537-9290-338b2c0a2d73\") " pod="openstack/placement-db-sync-slvbz" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.473810 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55091ec1-5a26-4825-9a61-a49cecda5a6f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"55091ec1-5a26-4825-9a61-a49cecda5a6f\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.473831 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55091ec1-5a26-4825-9a61-a49cecda5a6f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"55091ec1-5a26-4825-9a61-a49cecda5a6f\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.473847 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6af63266-547b-4537-9290-338b2c0a2d73-scripts\") pod \"placement-db-sync-slvbz\" (UID: \"6af63266-547b-4537-9290-338b2c0a2d73\") " pod="openstack/placement-db-sync-slvbz" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.473874 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9pxp\" (UniqueName: \"kubernetes.io/projected/6af63266-547b-4537-9290-338b2c0a2d73-kube-api-access-m9pxp\") pod \"placement-db-sync-slvbz\" (UID: \"6af63266-547b-4537-9290-338b2c0a2d73\") " pod="openstack/placement-db-sync-slvbz" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.473892 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.473914 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6af63266-547b-4537-9290-338b2c0a2d73-combined-ca-bundle\") pod \"placement-db-sync-slvbz\" (UID: \"6af63266-547b-4537-9290-338b2c0a2d73\") " pod="openstack/placement-db-sync-slvbz" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.473963 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/55091ec1-5a26-4825-9a61-a49cecda5a6f-logs\") pod \"glance-default-internal-api-0\" (UID: \"55091ec1-5a26-4825-9a61-a49cecda5a6f\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.473984 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3-config-data\") pod \"glance-default-external-api-0\" (UID: \"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.474003 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljbw8\" (UniqueName: \"kubernetes.io/projected/a63d659d-52a4-4f33-85c2-1461423776a3-kube-api-access-ljbw8\") pod \"dnsmasq-dns-7bb97fccb5-vblx7\" (UID: \"a63d659d-52a4-4f33-85c2-1461423776a3\") " pod="openstack/dnsmasq-dns-7bb97fccb5-vblx7" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.474015 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3-logs\") pod \"glance-default-external-api-0\" (UID: \"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.474026 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84s2j\" (UniqueName: \"kubernetes.io/projected/1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3-kube-api-access-84s2j\") pod \"glance-default-external-api-0\" (UID: \"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.474107 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6af63266-547b-4537-9290-338b2c0a2d73-logs\") pod \"placement-db-sync-slvbz\" (UID: \"6af63266-547b-4537-9290-338b2c0a2d73\") " pod="openstack/placement-db-sync-slvbz" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.474459 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6af63266-547b-4537-9290-338b2c0a2d73-logs\") pod \"placement-db-sync-slvbz\" (UID: \"6af63266-547b-4537-9290-338b2c0a2d73\") " pod="openstack/placement-db-sync-slvbz" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.475611 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.475964 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.479643 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3-scripts\") pod \"glance-default-external-api-0\" (UID: \"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.479911 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.483008 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6af63266-547b-4537-9290-338b2c0a2d73-config-data\") pod \"placement-db-sync-slvbz\" (UID: \"6af63266-547b-4537-9290-338b2c0a2d73\") " pod="openstack/placement-db-sync-slvbz" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.483030 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3-config-data\") pod \"glance-default-external-api-0\" (UID: \"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.484681 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6af63266-547b-4537-9290-338b2c0a2d73-combined-ca-bundle\") pod \"placement-db-sync-slvbz\" (UID: \"6af63266-547b-4537-9290-338b2c0a2d73\") " pod="openstack/placement-db-sync-slvbz" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.495515 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6af63266-547b-4537-9290-338b2c0a2d73-scripts\") pod \"placement-db-sync-slvbz\" (UID: \"6af63266-547b-4537-9290-338b2c0a2d73\") " pod="openstack/placement-db-sync-slvbz" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.499641 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9pxp\" (UniqueName: \"kubernetes.io/projected/6af63266-547b-4537-9290-338b2c0a2d73-kube-api-access-m9pxp\") pod \"placement-db-sync-slvbz\" (UID: \"6af63266-547b-4537-9290-338b2c0a2d73\") " pod="openstack/placement-db-sync-slvbz" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.503445 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84s2j\" (UniqueName: \"kubernetes.io/projected/1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3-kube-api-access-84s2j\") pod \"glance-default-external-api-0\" (UID: \"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.526434 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-b48666669-4sqql" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.535855 4787 generic.go:334] "Generic (PLEG): container finished" podID="80bd8ecf-a2fb-4635-83f1-050e227e74e6" containerID="02b4037afb172a39d39693d2e624c65b1326bc9b665ead63181fab2ff8180c5c" exitCode=0 Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.535937 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-l8fqd-config-jw2wk" event={"ID":"80bd8ecf-a2fb-4635-83f1-050e227e74e6","Type":"ContainerDied","Data":"02b4037afb172a39d39693d2e624c65b1326bc9b665ead63181fab2ff8180c5c"} Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.542565 4787 generic.go:334] "Generic (PLEG): container finished" podID="6754353d-92dc-4f0b-8e19-4ac10e7210d7" containerID="c48ff06cabd68d0b01fd21b7e9e29baae8ba587622ff9f69762494ed4d23f48d" exitCode=0 Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.542622 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4b6d4f-5l2dw" event={"ID":"6754353d-92dc-4f0b-8e19-4ac10e7210d7","Type":"ContainerDied","Data":"c48ff06cabd68d0b01fd21b7e9e29baae8ba587622ff9f69762494ed4d23f48d"} Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.542648 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4b6d4f-5l2dw" event={"ID":"6754353d-92dc-4f0b-8e19-4ac10e7210d7","Type":"ContainerStarted","Data":"7afd6cef81af428cf3d24012fcef177de1e8d5b57aad85ff41fe2e3448057adc"} Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.568266 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.577486 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a63d659d-52a4-4f33-85c2-1461423776a3-dns-svc\") pod \"dnsmasq-dns-7bb97fccb5-vblx7\" (UID: \"a63d659d-52a4-4f33-85c2-1461423776a3\") " pod="openstack/dnsmasq-dns-7bb97fccb5-vblx7" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.577532 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qllv9\" (UniqueName: \"kubernetes.io/projected/55091ec1-5a26-4825-9a61-a49cecda5a6f-kube-api-access-qllv9\") pod \"glance-default-internal-api-0\" (UID: \"55091ec1-5a26-4825-9a61-a49cecda5a6f\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.577555 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a63d659d-52a4-4f33-85c2-1461423776a3-config\") pod \"dnsmasq-dns-7bb97fccb5-vblx7\" (UID: \"a63d659d-52a4-4f33-85c2-1461423776a3\") " pod="openstack/dnsmasq-dns-7bb97fccb5-vblx7" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.577575 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a63d659d-52a4-4f33-85c2-1461423776a3-ovsdbserver-sb\") pod \"dnsmasq-dns-7bb97fccb5-vblx7\" (UID: \"a63d659d-52a4-4f33-85c2-1461423776a3\") " pod="openstack/dnsmasq-dns-7bb97fccb5-vblx7" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.577640 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a63d659d-52a4-4f33-85c2-1461423776a3-ovsdbserver-nb\") pod \"dnsmasq-dns-7bb97fccb5-vblx7\" (UID: \"a63d659d-52a4-4f33-85c2-1461423776a3\") " pod="openstack/dnsmasq-dns-7bb97fccb5-vblx7" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.577661 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/55091ec1-5a26-4825-9a61-a49cecda5a6f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"55091ec1-5a26-4825-9a61-a49cecda5a6f\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.577675 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55091ec1-5a26-4825-9a61-a49cecda5a6f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"55091ec1-5a26-4825-9a61-a49cecda5a6f\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.577697 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"55091ec1-5a26-4825-9a61-a49cecda5a6f\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.577720 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55091ec1-5a26-4825-9a61-a49cecda5a6f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"55091ec1-5a26-4825-9a61-a49cecda5a6f\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.577744 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55091ec1-5a26-4825-9a61-a49cecda5a6f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"55091ec1-5a26-4825-9a61-a49cecda5a6f\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.577829 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/55091ec1-5a26-4825-9a61-a49cecda5a6f-logs\") pod \"glance-default-internal-api-0\" (UID: \"55091ec1-5a26-4825-9a61-a49cecda5a6f\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.577858 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljbw8\" (UniqueName: \"kubernetes.io/projected/a63d659d-52a4-4f33-85c2-1461423776a3-kube-api-access-ljbw8\") pod \"dnsmasq-dns-7bb97fccb5-vblx7\" (UID: \"a63d659d-52a4-4f33-85c2-1461423776a3\") " pod="openstack/dnsmasq-dns-7bb97fccb5-vblx7" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.578577 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/55091ec1-5a26-4825-9a61-a49cecda5a6f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"55091ec1-5a26-4825-9a61-a49cecda5a6f\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.578649 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a63d659d-52a4-4f33-85c2-1461423776a3-ovsdbserver-nb\") pod \"dnsmasq-dns-7bb97fccb5-vblx7\" (UID: \"a63d659d-52a4-4f33-85c2-1461423776a3\") " pod="openstack/dnsmasq-dns-7bb97fccb5-vblx7" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.580647 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a63d659d-52a4-4f33-85c2-1461423776a3-dns-svc\") pod \"dnsmasq-dns-7bb97fccb5-vblx7\" (UID: \"a63d659d-52a4-4f33-85c2-1461423776a3\") " pod="openstack/dnsmasq-dns-7bb97fccb5-vblx7" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.581289 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a63d659d-52a4-4f33-85c2-1461423776a3-ovsdbserver-sb\") pod \"dnsmasq-dns-7bb97fccb5-vblx7\" (UID: \"a63d659d-52a4-4f33-85c2-1461423776a3\") " pod="openstack/dnsmasq-dns-7bb97fccb5-vblx7" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.581391 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a63d659d-52a4-4f33-85c2-1461423776a3-config\") pod \"dnsmasq-dns-7bb97fccb5-vblx7\" (UID: \"a63d659d-52a4-4f33-85c2-1461423776a3\") " pod="openstack/dnsmasq-dns-7bb97fccb5-vblx7" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.581571 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/55091ec1-5a26-4825-9a61-a49cecda5a6f-logs\") pod \"glance-default-internal-api-0\" (UID: \"55091ec1-5a26-4825-9a61-a49cecda5a6f\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.581665 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"55091ec1-5a26-4825-9a61-a49cecda5a6f\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.582710 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55091ec1-5a26-4825-9a61-a49cecda5a6f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"55091ec1-5a26-4825-9a61-a49cecda5a6f\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.585143 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55091ec1-5a26-4825-9a61-a49cecda5a6f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"55091ec1-5a26-4825-9a61-a49cecda5a6f\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.588004 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55091ec1-5a26-4825-9a61-a49cecda5a6f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"55091ec1-5a26-4825-9a61-a49cecda5a6f\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.606489 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljbw8\" (UniqueName: \"kubernetes.io/projected/a63d659d-52a4-4f33-85c2-1461423776a3-kube-api-access-ljbw8\") pod \"dnsmasq-dns-7bb97fccb5-vblx7\" (UID: \"a63d659d-52a4-4f33-85c2-1461423776a3\") " pod="openstack/dnsmasq-dns-7bb97fccb5-vblx7" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.606972 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qllv9\" (UniqueName: \"kubernetes.io/projected/55091ec1-5a26-4825-9a61-a49cecda5a6f-kube-api-access-qllv9\") pod \"glance-default-internal-api-0\" (UID: \"55091ec1-5a26-4825-9a61-a49cecda5a6f\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.607347 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-slvbz" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.617244 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"55091ec1-5a26-4825-9a61-a49cecda5a6f\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.625677 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.695366 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bb97fccb5-vblx7" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:01.767533 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:02.411859 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4b6d4f-5l2dw" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:02.552053 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4b6d4f-5l2dw" event={"ID":"6754353d-92dc-4f0b-8e19-4ac10e7210d7","Type":"ContainerDied","Data":"7afd6cef81af428cf3d24012fcef177de1e8d5b57aad85ff41fe2e3448057adc"} Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:02.552066 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4b6d4f-5l2dw" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:02.552130 4787 scope.go:117] "RemoveContainer" containerID="c48ff06cabd68d0b01fd21b7e9e29baae8ba587622ff9f69762494ed4d23f48d" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:02.598266 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xz2fw\" (UniqueName: \"kubernetes.io/projected/6754353d-92dc-4f0b-8e19-4ac10e7210d7-kube-api-access-xz2fw\") pod \"6754353d-92dc-4f0b-8e19-4ac10e7210d7\" (UID: \"6754353d-92dc-4f0b-8e19-4ac10e7210d7\") " Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:02.598437 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6754353d-92dc-4f0b-8e19-4ac10e7210d7-ovsdbserver-nb\") pod \"6754353d-92dc-4f0b-8e19-4ac10e7210d7\" (UID: \"6754353d-92dc-4f0b-8e19-4ac10e7210d7\") " Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:02.598470 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6754353d-92dc-4f0b-8e19-4ac10e7210d7-config\") pod \"6754353d-92dc-4f0b-8e19-4ac10e7210d7\" (UID: \"6754353d-92dc-4f0b-8e19-4ac10e7210d7\") " Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:02.598636 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6754353d-92dc-4f0b-8e19-4ac10e7210d7-dns-svc\") pod \"6754353d-92dc-4f0b-8e19-4ac10e7210d7\" (UID: \"6754353d-92dc-4f0b-8e19-4ac10e7210d7\") " Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:02.604825 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6754353d-92dc-4f0b-8e19-4ac10e7210d7-ovsdbserver-sb\") pod \"6754353d-92dc-4f0b-8e19-4ac10e7210d7\" (UID: \"6754353d-92dc-4f0b-8e19-4ac10e7210d7\") " Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:02.610917 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6754353d-92dc-4f0b-8e19-4ac10e7210d7-kube-api-access-xz2fw" (OuterVolumeSpecName: "kube-api-access-xz2fw") pod "6754353d-92dc-4f0b-8e19-4ac10e7210d7" (UID: "6754353d-92dc-4f0b-8e19-4ac10e7210d7"). InnerVolumeSpecName "kube-api-access-xz2fw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:02.626939 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6754353d-92dc-4f0b-8e19-4ac10e7210d7-config" (OuterVolumeSpecName: "config") pod "6754353d-92dc-4f0b-8e19-4ac10e7210d7" (UID: "6754353d-92dc-4f0b-8e19-4ac10e7210d7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:02.631597 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6754353d-92dc-4f0b-8e19-4ac10e7210d7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6754353d-92dc-4f0b-8e19-4ac10e7210d7" (UID: "6754353d-92dc-4f0b-8e19-4ac10e7210d7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:02.632387 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6754353d-92dc-4f0b-8e19-4ac10e7210d7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6754353d-92dc-4f0b-8e19-4ac10e7210d7" (UID: "6754353d-92dc-4f0b-8e19-4ac10e7210d7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:02.645538 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6754353d-92dc-4f0b-8e19-4ac10e7210d7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6754353d-92dc-4f0b-8e19-4ac10e7210d7" (UID: "6754353d-92dc-4f0b-8e19-4ac10e7210d7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:02.707908 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xz2fw\" (UniqueName: \"kubernetes.io/projected/6754353d-92dc-4f0b-8e19-4ac10e7210d7-kube-api-access-xz2fw\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:02.707946 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6754353d-92dc-4f0b-8e19-4ac10e7210d7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:02.707955 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6754353d-92dc-4f0b-8e19-4ac10e7210d7-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:02.707965 4787 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6754353d-92dc-4f0b-8e19-4ac10e7210d7-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:02.707973 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6754353d-92dc-4f0b-8e19-4ac10e7210d7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:02.733739 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-58779b6b6f-rqbn7"] Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:02.748987 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-slvbz"] Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:02.759428 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cd86cbb59-nqs8t"] Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:02.773149 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-b48666669-4sqql"] Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:02.793979 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:02.808247 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-tk8w8"] Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:02.818563 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bb97fccb5-vblx7"] Oct 01 09:52:02 crc kubenswrapper[4787]: W1001 09:52:02.831771 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9587d64e_33e3_45d2_8ee7_a776fcc60d88.slice/crio-3a332e3f42bdb96a3a7da96a0f57511b9d0dde1414884f7ff47a9d91125a8639 WatchSource:0}: Error finding container 3a332e3f42bdb96a3a7da96a0f57511b9d0dde1414884f7ff47a9d91125a8639: Status 404 returned error can't find the container with id 3a332e3f42bdb96a3a7da96a0f57511b9d0dde1414884f7ff47a9d91125a8639 Oct 01 09:52:02 crc kubenswrapper[4787]: I1001 09:52:02.922685 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.188500 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-l8fqd-config-jw2wk" Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.217384 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/80bd8ecf-a2fb-4635-83f1-050e227e74e6-scripts\") pod \"80bd8ecf-a2fb-4635-83f1-050e227e74e6\" (UID: \"80bd8ecf-a2fb-4635-83f1-050e227e74e6\") " Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.217476 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26wbl\" (UniqueName: \"kubernetes.io/projected/80bd8ecf-a2fb-4635-83f1-050e227e74e6-kube-api-access-26wbl\") pod \"80bd8ecf-a2fb-4635-83f1-050e227e74e6\" (UID: \"80bd8ecf-a2fb-4635-83f1-050e227e74e6\") " Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.217497 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/80bd8ecf-a2fb-4635-83f1-050e227e74e6-var-run-ovn\") pod \"80bd8ecf-a2fb-4635-83f1-050e227e74e6\" (UID: \"80bd8ecf-a2fb-4635-83f1-050e227e74e6\") " Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.217517 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/80bd8ecf-a2fb-4635-83f1-050e227e74e6-var-run\") pod \"80bd8ecf-a2fb-4635-83f1-050e227e74e6\" (UID: \"80bd8ecf-a2fb-4635-83f1-050e227e74e6\") " Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.217599 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/80bd8ecf-a2fb-4635-83f1-050e227e74e6-var-log-ovn\") pod \"80bd8ecf-a2fb-4635-83f1-050e227e74e6\" (UID: \"80bd8ecf-a2fb-4635-83f1-050e227e74e6\") " Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.217662 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/80bd8ecf-a2fb-4635-83f1-050e227e74e6-additional-scripts\") pod \"80bd8ecf-a2fb-4635-83f1-050e227e74e6\" (UID: \"80bd8ecf-a2fb-4635-83f1-050e227e74e6\") " Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.218200 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/80bd8ecf-a2fb-4635-83f1-050e227e74e6-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "80bd8ecf-a2fb-4635-83f1-050e227e74e6" (UID: "80bd8ecf-a2fb-4635-83f1-050e227e74e6"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.218271 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/80bd8ecf-a2fb-4635-83f1-050e227e74e6-var-run" (OuterVolumeSpecName: "var-run") pod "80bd8ecf-a2fb-4635-83f1-050e227e74e6" (UID: "80bd8ecf-a2fb-4635-83f1-050e227e74e6"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.218570 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/80bd8ecf-a2fb-4635-83f1-050e227e74e6-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "80bd8ecf-a2fb-4635-83f1-050e227e74e6" (UID: "80bd8ecf-a2fb-4635-83f1-050e227e74e6"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.219480 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80bd8ecf-a2fb-4635-83f1-050e227e74e6-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "80bd8ecf-a2fb-4635-83f1-050e227e74e6" (UID: "80bd8ecf-a2fb-4635-83f1-050e227e74e6"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.219733 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80bd8ecf-a2fb-4635-83f1-050e227e74e6-scripts" (OuterVolumeSpecName: "scripts") pod "80bd8ecf-a2fb-4635-83f1-050e227e74e6" (UID: "80bd8ecf-a2fb-4635-83f1-050e227e74e6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.230596 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80bd8ecf-a2fb-4635-83f1-050e227e74e6-kube-api-access-26wbl" (OuterVolumeSpecName: "kube-api-access-26wbl") pod "80bd8ecf-a2fb-4635-83f1-050e227e74e6" (UID: "80bd8ecf-a2fb-4635-83f1-050e227e74e6"). InnerVolumeSpecName "kube-api-access-26wbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.275502 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4b6d4f-5l2dw"] Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.287729 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-757b4b6d4f-5l2dw"] Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.320345 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/80bd8ecf-a2fb-4635-83f1-050e227e74e6-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.320517 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26wbl\" (UniqueName: \"kubernetes.io/projected/80bd8ecf-a2fb-4635-83f1-050e227e74e6-kube-api-access-26wbl\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.320571 4787 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/80bd8ecf-a2fb-4635-83f1-050e227e74e6-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.320619 4787 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/80bd8ecf-a2fb-4635-83f1-050e227e74e6-var-run\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.320682 4787 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/80bd8ecf-a2fb-4635-83f1-050e227e74e6-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.320756 4787 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/80bd8ecf-a2fb-4635-83f1-050e227e74e6-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.601731 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.625396 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-slvbz" event={"ID":"6af63266-547b-4537-9290-338b2c0a2d73","Type":"ContainerStarted","Data":"b8ec9efd63e8d65d90b0fb7729f20b38142a574a75c8a295eb585b2029d2b07c"} Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.646371 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-l8fqd-config-jw2wk" Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.647277 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-l8fqd-config-jw2wk" event={"ID":"80bd8ecf-a2fb-4635-83f1-050e227e74e6","Type":"ContainerDied","Data":"5afc08c7207c715e1f38eebf786d7ad575fa03c0ea2fcada177b81cca344542e"} Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.647341 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5afc08c7207c715e1f38eebf786d7ad575fa03c0ea2fcada177b81cca344542e" Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.671041 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-b48666669-4sqql" event={"ID":"3869928b-1191-433d-87bf-7187615b57e0","Type":"ContainerStarted","Data":"826baf402d328a062c18d66aeae1e510c3b2433b269b21aa1da7daceb12f0b51"} Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.672654 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3","Type":"ContainerStarted","Data":"9900a4add31c368beea14cfcefbd6807b21c8275eb21edd9ea8ae4f00d6d8964"} Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.674131 4787 generic.go:334] "Generic (PLEG): container finished" podID="9c3e6671-985a-4fc7-a8a0-17f290f6a382" containerID="e8215c2de8f1c5bb21ada2437695e36fc0cd75c43bc5dcc5ff2b488f8c7fdd57" exitCode=0 Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.674191 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cd86cbb59-nqs8t" event={"ID":"9c3e6671-985a-4fc7-a8a0-17f290f6a382","Type":"ContainerDied","Data":"e8215c2de8f1c5bb21ada2437695e36fc0cd75c43bc5dcc5ff2b488f8c7fdd57"} Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.674211 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cd86cbb59-nqs8t" event={"ID":"9c3e6671-985a-4fc7-a8a0-17f290f6a382","Type":"ContainerStarted","Data":"f4204ac4104c9b911c45a67d31502c4e49bc7df3bb027e560f922ef0b1accd18"} Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.699413 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-tk8w8" event={"ID":"266f85d9-996e-41a9-a83d-5451e74689b6","Type":"ContainerStarted","Data":"48a9a1fbf4d60825c9724f9ea3da0e4ff19da26f8e5bac0bf0f259bd88d689bf"} Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.699486 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-tk8w8" event={"ID":"266f85d9-996e-41a9-a83d-5451e74689b6","Type":"ContainerStarted","Data":"42049a54e6439475e05e6c3b439358d6220f361271336081ec8f5793e3599625"} Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.702289 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.712908 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9587d64e-33e3-45d2-8ee7-a776fcc60d88","Type":"ContainerStarted","Data":"3a332e3f42bdb96a3a7da96a0f57511b9d0dde1414884f7ff47a9d91125a8639"} Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.742057 4787 generic.go:334] "Generic (PLEG): container finished" podID="a63d659d-52a4-4f33-85c2-1461423776a3" containerID="b4b7dcf91d6035d9dc19d6070186f3ddafa3f9f0b196a665c5a018cd70250c71" exitCode=0 Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.742178 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bb97fccb5-vblx7" event={"ID":"a63d659d-52a4-4f33-85c2-1461423776a3","Type":"ContainerDied","Data":"b4b7dcf91d6035d9dc19d6070186f3ddafa3f9f0b196a665c5a018cd70250c71"} Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.742204 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bb97fccb5-vblx7" event={"ID":"a63d659d-52a4-4f33-85c2-1461423776a3","Type":"ContainerStarted","Data":"acfe6f4935f074301a7a9f7a2bd89dfb417aa356089cdb784bda91e6255a9104"} Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.776745 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.798283 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-58779b6b6f-rqbn7" event={"ID":"2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3","Type":"ContainerStarted","Data":"cea04ca942277b341a7475576791e4622eb8bc4d8d205d7cbc5befacc79d6de3"} Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.816139 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-58779b6b6f-rqbn7"] Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.861578 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7565dcfdb5-mcrm8"] Oct 01 09:52:03 crc kubenswrapper[4787]: E1001 09:52:03.862066 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80bd8ecf-a2fb-4635-83f1-050e227e74e6" containerName="ovn-config" Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.862100 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="80bd8ecf-a2fb-4635-83f1-050e227e74e6" containerName="ovn-config" Oct 01 09:52:03 crc kubenswrapper[4787]: E1001 09:52:03.862134 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6754353d-92dc-4f0b-8e19-4ac10e7210d7" containerName="init" Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.862141 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="6754353d-92dc-4f0b-8e19-4ac10e7210d7" containerName="init" Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.862363 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="80bd8ecf-a2fb-4635-83f1-050e227e74e6" containerName="ovn-config" Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.862385 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="6754353d-92dc-4f0b-8e19-4ac10e7210d7" containerName="init" Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.867768 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7565dcfdb5-mcrm8" Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.887187 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.899560 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-tk8w8" podStartSLOduration=3.899535203 podStartE2EDuration="3.899535203s" podCreationTimestamp="2025-10-01 09:52:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:52:03.856869259 +0000 UTC m=+955.972013416" watchObservedRunningTime="2025-10-01 09:52:03.899535203 +0000 UTC m=+956.014679360" Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.925312 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7565dcfdb5-mcrm8"] Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.942466 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/54c87b91-944a-4f8c-a556-21649f3b9b1b-scripts\") pod \"horizon-7565dcfdb5-mcrm8\" (UID: \"54c87b91-944a-4f8c-a556-21649f3b9b1b\") " pod="openstack/horizon-7565dcfdb5-mcrm8" Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.942545 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/54c87b91-944a-4f8c-a556-21649f3b9b1b-logs\") pod \"horizon-7565dcfdb5-mcrm8\" (UID: \"54c87b91-944a-4f8c-a556-21649f3b9b1b\") " pod="openstack/horizon-7565dcfdb5-mcrm8" Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.942610 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/54c87b91-944a-4f8c-a556-21649f3b9b1b-horizon-secret-key\") pod \"horizon-7565dcfdb5-mcrm8\" (UID: \"54c87b91-944a-4f8c-a556-21649f3b9b1b\") " pod="openstack/horizon-7565dcfdb5-mcrm8" Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.942650 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdpd2\" (UniqueName: \"kubernetes.io/projected/54c87b91-944a-4f8c-a556-21649f3b9b1b-kube-api-access-mdpd2\") pod \"horizon-7565dcfdb5-mcrm8\" (UID: \"54c87b91-944a-4f8c-a556-21649f3b9b1b\") " pod="openstack/horizon-7565dcfdb5-mcrm8" Oct 01 09:52:03 crc kubenswrapper[4787]: I1001 09:52:03.942731 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/54c87b91-944a-4f8c-a556-21649f3b9b1b-config-data\") pod \"horizon-7565dcfdb5-mcrm8\" (UID: \"54c87b91-944a-4f8c-a556-21649f3b9b1b\") " pod="openstack/horizon-7565dcfdb5-mcrm8" Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.050523 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/54c87b91-944a-4f8c-a556-21649f3b9b1b-scripts\") pod \"horizon-7565dcfdb5-mcrm8\" (UID: \"54c87b91-944a-4f8c-a556-21649f3b9b1b\") " pod="openstack/horizon-7565dcfdb5-mcrm8" Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.050569 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/54c87b91-944a-4f8c-a556-21649f3b9b1b-logs\") pod \"horizon-7565dcfdb5-mcrm8\" (UID: \"54c87b91-944a-4f8c-a556-21649f3b9b1b\") " pod="openstack/horizon-7565dcfdb5-mcrm8" Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.050596 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/54c87b91-944a-4f8c-a556-21649f3b9b1b-horizon-secret-key\") pod \"horizon-7565dcfdb5-mcrm8\" (UID: \"54c87b91-944a-4f8c-a556-21649f3b9b1b\") " pod="openstack/horizon-7565dcfdb5-mcrm8" Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.050624 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdpd2\" (UniqueName: \"kubernetes.io/projected/54c87b91-944a-4f8c-a556-21649f3b9b1b-kube-api-access-mdpd2\") pod \"horizon-7565dcfdb5-mcrm8\" (UID: \"54c87b91-944a-4f8c-a556-21649f3b9b1b\") " pod="openstack/horizon-7565dcfdb5-mcrm8" Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.050664 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/54c87b91-944a-4f8c-a556-21649f3b9b1b-config-data\") pod \"horizon-7565dcfdb5-mcrm8\" (UID: \"54c87b91-944a-4f8c-a556-21649f3b9b1b\") " pod="openstack/horizon-7565dcfdb5-mcrm8" Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.051751 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/54c87b91-944a-4f8c-a556-21649f3b9b1b-logs\") pod \"horizon-7565dcfdb5-mcrm8\" (UID: \"54c87b91-944a-4f8c-a556-21649f3b9b1b\") " pod="openstack/horizon-7565dcfdb5-mcrm8" Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.052030 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/54c87b91-944a-4f8c-a556-21649f3b9b1b-config-data\") pod \"horizon-7565dcfdb5-mcrm8\" (UID: \"54c87b91-944a-4f8c-a556-21649f3b9b1b\") " pod="openstack/horizon-7565dcfdb5-mcrm8" Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.052593 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/54c87b91-944a-4f8c-a556-21649f3b9b1b-scripts\") pod \"horizon-7565dcfdb5-mcrm8\" (UID: \"54c87b91-944a-4f8c-a556-21649f3b9b1b\") " pod="openstack/horizon-7565dcfdb5-mcrm8" Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.066351 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/54c87b91-944a-4f8c-a556-21649f3b9b1b-horizon-secret-key\") pod \"horizon-7565dcfdb5-mcrm8\" (UID: \"54c87b91-944a-4f8c-a556-21649f3b9b1b\") " pod="openstack/horizon-7565dcfdb5-mcrm8" Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.080754 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdpd2\" (UniqueName: \"kubernetes.io/projected/54c87b91-944a-4f8c-a556-21649f3b9b1b-kube-api-access-mdpd2\") pod \"horizon-7565dcfdb5-mcrm8\" (UID: \"54c87b91-944a-4f8c-a556-21649f3b9b1b\") " pod="openstack/horizon-7565dcfdb5-mcrm8" Oct 01 09:52:04 crc kubenswrapper[4787]: E1001 09:52:04.202373 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.io/podified-antelope-centos9/openstack-swift-object@sha256:793a836e17b07b0e0a4e8d3177fd04724e1e058fca275ef434abe60a2e444a79: can't talk to a V1 container registry" image="quay.io/podified-antelope-centos9/openstack-swift-object@sha256:793a836e17b07b0e0a4e8d3177fd04724e1e058fca275ef434abe60a2e444a79" Oct 01 09:52:04 crc kubenswrapper[4787]: E1001 09:52:04.202645 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:object-server,Image:quay.io/podified-antelope-centos9/openstack-swift-object@sha256:793a836e17b07b0e0a4e8d3177fd04724e1e058fca275ef434abe60a2e444a79,Command:[/usr/bin/swift-object-server /etc/swift/object-server.conf.d -v],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:object,HostPort:0,ContainerPort:6200,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5b7h56h9dh94h67bh697h95h55hbh555h556h675h5fdh57dh579h5fbh64fh5c9h687hb6h678h5d4h549h54h98h8ch564h5bh5bch55dhc8hf8q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:swift,ReadOnly:false,MountPath:/srv/node/pv,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:etc-swift,ReadOnly:false,MountPath:/etc/swift,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cache,ReadOnly:false,MountPath:/var/cache/swift,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:lock,ReadOnly:false,MountPath:/var/lock,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vzdqq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42445,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-storage-0_openstack(3432bc16-99b4-431a-aeb1-600e826bbc3e): ErrImagePull: initializing source docker://quay.io/podified-antelope-centos9/openstack-swift-object@sha256:793a836e17b07b0e0a4e8d3177fd04724e1e058fca275ef434abe60a2e444a79: can't talk to a V1 container registry" logger="UnhandledError" Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.274495 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-l8fqd-config-jw2wk"] Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.287388 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-l8fqd-config-jw2wk"] Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.309573 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7565dcfdb5-mcrm8" Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.333192 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cd86cbb59-nqs8t" Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.361401 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9c3e6671-985a-4fc7-a8a0-17f290f6a382-ovsdbserver-nb\") pod \"9c3e6671-985a-4fc7-a8a0-17f290f6a382\" (UID: \"9c3e6671-985a-4fc7-a8a0-17f290f6a382\") " Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.361493 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9c3e6671-985a-4fc7-a8a0-17f290f6a382-ovsdbserver-sb\") pod \"9c3e6671-985a-4fc7-a8a0-17f290f6a382\" (UID: \"9c3e6671-985a-4fc7-a8a0-17f290f6a382\") " Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.361677 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c3e6671-985a-4fc7-a8a0-17f290f6a382-config\") pod \"9c3e6671-985a-4fc7-a8a0-17f290f6a382\" (UID: \"9c3e6671-985a-4fc7-a8a0-17f290f6a382\") " Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.361730 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5nhtc\" (UniqueName: \"kubernetes.io/projected/9c3e6671-985a-4fc7-a8a0-17f290f6a382-kube-api-access-5nhtc\") pod \"9c3e6671-985a-4fc7-a8a0-17f290f6a382\" (UID: \"9c3e6671-985a-4fc7-a8a0-17f290f6a382\") " Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.361761 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9c3e6671-985a-4fc7-a8a0-17f290f6a382-dns-svc\") pod \"9c3e6671-985a-4fc7-a8a0-17f290f6a382\" (UID: \"9c3e6671-985a-4fc7-a8a0-17f290f6a382\") " Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.391273 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c3e6671-985a-4fc7-a8a0-17f290f6a382-kube-api-access-5nhtc" (OuterVolumeSpecName: "kube-api-access-5nhtc") pod "9c3e6671-985a-4fc7-a8a0-17f290f6a382" (UID: "9c3e6671-985a-4fc7-a8a0-17f290f6a382"). InnerVolumeSpecName "kube-api-access-5nhtc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.404356 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c3e6671-985a-4fc7-a8a0-17f290f6a382-config" (OuterVolumeSpecName: "config") pod "9c3e6671-985a-4fc7-a8a0-17f290f6a382" (UID: "9c3e6671-985a-4fc7-a8a0-17f290f6a382"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.410543 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c3e6671-985a-4fc7-a8a0-17f290f6a382-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9c3e6671-985a-4fc7-a8a0-17f290f6a382" (UID: "9c3e6671-985a-4fc7-a8a0-17f290f6a382"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.413782 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c3e6671-985a-4fc7-a8a0-17f290f6a382-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9c3e6671-985a-4fc7-a8a0-17f290f6a382" (UID: "9c3e6671-985a-4fc7-a8a0-17f290f6a382"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.414291 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c3e6671-985a-4fc7-a8a0-17f290f6a382-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9c3e6671-985a-4fc7-a8a0-17f290f6a382" (UID: "9c3e6671-985a-4fc7-a8a0-17f290f6a382"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.516807 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9c3e6671-985a-4fc7-a8a0-17f290f6a382-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.516839 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9c3e6671-985a-4fc7-a8a0-17f290f6a382-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.516848 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c3e6671-985a-4fc7-a8a0-17f290f6a382-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.516857 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5nhtc\" (UniqueName: \"kubernetes.io/projected/9c3e6671-985a-4fc7-a8a0-17f290f6a382-kube-api-access-5nhtc\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.516889 4787 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9c3e6671-985a-4fc7-a8a0-17f290f6a382-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.550895 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6754353d-92dc-4f0b-8e19-4ac10e7210d7" path="/var/lib/kubelet/pods/6754353d-92dc-4f0b-8e19-4ac10e7210d7/volumes" Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.551668 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80bd8ecf-a2fb-4635-83f1-050e227e74e6" path="/var/lib/kubelet/pods/80bd8ecf-a2fb-4635-83f1-050e227e74e6/volumes" Oct 01 09:52:04 crc kubenswrapper[4787]: E1001 09:52:04.607152 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"container-server\" with ErrImagePull: \"initializing source docker://quay.io/podified-antelope-centos9/openstack-swift-container@sha256:7118cc3a695fead2a8bab14c8ace018ed7a5ba23ef347bf4ead44219e8467866: reading manifest sha256:7118cc3a695fead2a8bab14c8ace018ed7a5ba23ef347bf4ead44219e8467866 in quay.io/podified-antelope-centos9/openstack-swift-container: received unexpected HTTP status: 504 Gateway Time-out\", failed to \"StartContainer\" for \"container-replicator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-swift-container@sha256:7118cc3a695fead2a8bab14c8ace018ed7a5ba23ef347bf4ead44219e8467866\\\"\", failed to \"StartContainer\" for \"container-auditor\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-swift-container@sha256:7118cc3a695fead2a8bab14c8ace018ed7a5ba23ef347bf4ead44219e8467866\\\"\", failed to \"StartContainer\" for \"container-updater\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-swift-container@sha256:7118cc3a695fead2a8bab14c8ace018ed7a5ba23ef347bf4ead44219e8467866\\\"\", failed to \"StartContainer\" for \"object-server\" with ErrImagePull: \"initializing source docker://quay.io/podified-antelope-centos9/openstack-swift-object@sha256:793a836e17b07b0e0a4e8d3177fd04724e1e058fca275ef434abe60a2e444a79: can't talk to a V1 container registry\", failed to \"StartContainer\" for \"object-replicator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-swift-object@sha256:793a836e17b07b0e0a4e8d3177fd04724e1e058fca275ef434abe60a2e444a79\\\"\", failed to \"StartContainer\" for \"object-auditor\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-swift-object@sha256:793a836e17b07b0e0a4e8d3177fd04724e1e058fca275ef434abe60a2e444a79\\\"\", failed to \"StartContainer\" for \"object-updater\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-swift-object@sha256:793a836e17b07b0e0a4e8d3177fd04724e1e058fca275ef434abe60a2e444a79\\\"\", failed to \"StartContainer\" for \"rsync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-swift-object@sha256:793a836e17b07b0e0a4e8d3177fd04724e1e058fca275ef434abe60a2e444a79\\\"\", failed to \"StartContainer\" for \"swift-recon-cron\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-swift-object@sha256:793a836e17b07b0e0a4e8d3177fd04724e1e058fca275ef434abe60a2e444a79\\\"\"]" pod="openstack/swift-storage-0" podUID="3432bc16-99b4-431a-aeb1-600e826bbc3e" Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.824397 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"55091ec1-5a26-4825-9a61-a49cecda5a6f","Type":"ContainerStarted","Data":"ee2f77af998f7916b9be643dcc9322d0ea37a6c82778d6596e66bb6ad73b5c7f"} Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.840323 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3432bc16-99b4-431a-aeb1-600e826bbc3e","Type":"ContainerStarted","Data":"f3e59dd57873da3b0d68fa26207662846d4f5e6743ce0532a74823dd82351716"} Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.846826 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bb97fccb5-vblx7" event={"ID":"a63d659d-52a4-4f33-85c2-1461423776a3","Type":"ContainerStarted","Data":"4e7e606d8da369890eb621b428c3ac2e5da4b3dfcc45f7505cf0efc6d81bfb76"} Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.846896 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7bb97fccb5-vblx7" Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.850342 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3","Type":"ContainerStarted","Data":"bbd0a2d96558ec05c6a18fc511729c45b4afffab39efef62e9345bc1144c0d8f"} Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.853031 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cd86cbb59-nqs8t" Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.853540 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cd86cbb59-nqs8t" event={"ID":"9c3e6671-985a-4fc7-a8a0-17f290f6a382","Type":"ContainerDied","Data":"f4204ac4104c9b911c45a67d31502c4e49bc7df3bb027e560f922ef0b1accd18"} Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.853565 4787 scope.go:117] "RemoveContainer" containerID="e8215c2de8f1c5bb21ada2437695e36fc0cd75c43bc5dcc5ff2b488f8c7fdd57" Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.908166 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7565dcfdb5-mcrm8"] Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.981740 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cd86cbb59-nqs8t"] Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.989364 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7cd86cbb59-nqs8t"] Oct 01 09:52:04 crc kubenswrapper[4787]: I1001 09:52:04.990066 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7bb97fccb5-vblx7" podStartSLOduration=3.990049748 podStartE2EDuration="3.990049748s" podCreationTimestamp="2025-10-01 09:52:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:52:04.960348557 +0000 UTC m=+957.075492714" watchObservedRunningTime="2025-10-01 09:52:04.990049748 +0000 UTC m=+957.105193905" Oct 01 09:52:05 crc kubenswrapper[4787]: I1001 09:52:05.887543 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3","Type":"ContainerStarted","Data":"16d8a5275751e42e56226fe37a99ccde7f83e12a427fc146cbb0a8833077315c"} Oct 01 09:52:05 crc kubenswrapper[4787]: I1001 09:52:05.887911 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3" containerName="glance-log" containerID="cri-o://bbd0a2d96558ec05c6a18fc511729c45b4afffab39efef62e9345bc1144c0d8f" gracePeriod=30 Oct 01 09:52:05 crc kubenswrapper[4787]: I1001 09:52:05.888325 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3" containerName="glance-httpd" containerID="cri-o://16d8a5275751e42e56226fe37a99ccde7f83e12a427fc146cbb0a8833077315c" gracePeriod=30 Oct 01 09:52:05 crc kubenswrapper[4787]: I1001 09:52:05.896731 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"55091ec1-5a26-4825-9a61-a49cecda5a6f","Type":"ContainerStarted","Data":"0256ce9b0aaf2dc94f1c6ed17241671d2db6b17016086397971f59f0401d3e93"} Oct 01 09:52:05 crc kubenswrapper[4787]: I1001 09:52:05.896791 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="55091ec1-5a26-4825-9a61-a49cecda5a6f" containerName="glance-log" containerID="cri-o://c980e4cea0aa90172278d8133be1046b80fde5491d7e385d078843d35aa6b323" gracePeriod=30 Oct 01 09:52:05 crc kubenswrapper[4787]: I1001 09:52:05.896815 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"55091ec1-5a26-4825-9a61-a49cecda5a6f","Type":"ContainerStarted","Data":"c980e4cea0aa90172278d8133be1046b80fde5491d7e385d078843d35aa6b323"} Oct 01 09:52:05 crc kubenswrapper[4787]: I1001 09:52:05.896871 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="55091ec1-5a26-4825-9a61-a49cecda5a6f" containerName="glance-httpd" containerID="cri-o://0256ce9b0aaf2dc94f1c6ed17241671d2db6b17016086397971f59f0401d3e93" gracePeriod=30 Oct 01 09:52:05 crc kubenswrapper[4787]: I1001 09:52:05.909298 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7565dcfdb5-mcrm8" event={"ID":"54c87b91-944a-4f8c-a556-21649f3b9b1b","Type":"ContainerStarted","Data":"8fe9847b510592d21529b6e9e707acfc71ff4f1873c32df5efe1b30aa84f7c07"} Oct 01 09:52:05 crc kubenswrapper[4787]: I1001 09:52:05.921753 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.921726665 podStartE2EDuration="4.921726665s" podCreationTimestamp="2025-10-01 09:52:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:52:05.909590313 +0000 UTC m=+958.024734460" watchObservedRunningTime="2025-10-01 09:52:05.921726665 +0000 UTC m=+958.036870822" Oct 01 09:52:05 crc kubenswrapper[4787]: I1001 09:52:05.935634 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.935614911 podStartE2EDuration="4.935614911s" podCreationTimestamp="2025-10-01 09:52:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:52:05.929626708 +0000 UTC m=+958.044770875" watchObservedRunningTime="2025-10-01 09:52:05.935614911 +0000 UTC m=+958.050759068" Oct 01 09:52:06 crc kubenswrapper[4787]: I1001 09:52:06.547912 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c3e6671-985a-4fc7-a8a0-17f290f6a382" path="/var/lib/kubelet/pods/9c3e6671-985a-4fc7-a8a0-17f290f6a382/volumes" Oct 01 09:52:06 crc kubenswrapper[4787]: I1001 09:52:06.922996 4787 generic.go:334] "Generic (PLEG): container finished" podID="1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3" containerID="16d8a5275751e42e56226fe37a99ccde7f83e12a427fc146cbb0a8833077315c" exitCode=0 Oct 01 09:52:06 crc kubenswrapper[4787]: I1001 09:52:06.923030 4787 generic.go:334] "Generic (PLEG): container finished" podID="1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3" containerID="bbd0a2d96558ec05c6a18fc511729c45b4afffab39efef62e9345bc1144c0d8f" exitCode=143 Oct 01 09:52:06 crc kubenswrapper[4787]: I1001 09:52:06.923045 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3","Type":"ContainerDied","Data":"16d8a5275751e42e56226fe37a99ccde7f83e12a427fc146cbb0a8833077315c"} Oct 01 09:52:06 crc kubenswrapper[4787]: I1001 09:52:06.923114 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3","Type":"ContainerDied","Data":"bbd0a2d96558ec05c6a18fc511729c45b4afffab39efef62e9345bc1144c0d8f"} Oct 01 09:52:06 crc kubenswrapper[4787]: I1001 09:52:06.933949 4787 generic.go:334] "Generic (PLEG): container finished" podID="55091ec1-5a26-4825-9a61-a49cecda5a6f" containerID="0256ce9b0aaf2dc94f1c6ed17241671d2db6b17016086397971f59f0401d3e93" exitCode=143 Oct 01 09:52:06 crc kubenswrapper[4787]: I1001 09:52:06.933989 4787 generic.go:334] "Generic (PLEG): container finished" podID="55091ec1-5a26-4825-9a61-a49cecda5a6f" containerID="c980e4cea0aa90172278d8133be1046b80fde5491d7e385d078843d35aa6b323" exitCode=143 Oct 01 09:52:06 crc kubenswrapper[4787]: I1001 09:52:06.934013 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"55091ec1-5a26-4825-9a61-a49cecda5a6f","Type":"ContainerDied","Data":"0256ce9b0aaf2dc94f1c6ed17241671d2db6b17016086397971f59f0401d3e93"} Oct 01 09:52:06 crc kubenswrapper[4787]: I1001 09:52:06.934040 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"55091ec1-5a26-4825-9a61-a49cecda5a6f","Type":"ContainerDied","Data":"c980e4cea0aa90172278d8133be1046b80fde5491d7e385d078843d35aa6b323"} Oct 01 09:52:07 crc kubenswrapper[4787]: I1001 09:52:07.952692 4787 generic.go:334] "Generic (PLEG): container finished" podID="266f85d9-996e-41a9-a83d-5451e74689b6" containerID="48a9a1fbf4d60825c9724f9ea3da0e4ff19da26f8e5bac0bf0f259bd88d689bf" exitCode=0 Oct 01 09:52:07 crc kubenswrapper[4787]: I1001 09:52:07.952776 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-tk8w8" event={"ID":"266f85d9-996e-41a9-a83d-5451e74689b6","Type":"ContainerDied","Data":"48a9a1fbf4d60825c9724f9ea3da0e4ff19da26f8e5bac0bf0f259bd88d689bf"} Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.318883 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-eafd-account-create-h97jm"] Oct 01 09:52:08 crc kubenswrapper[4787]: E1001 09:52:08.319437 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c3e6671-985a-4fc7-a8a0-17f290f6a382" containerName="init" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.319467 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c3e6671-985a-4fc7-a8a0-17f290f6a382" containerName="init" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.319693 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c3e6671-985a-4fc7-a8a0-17f290f6a382" containerName="init" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.323929 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-eafd-account-create-h97jm" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.326775 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.331417 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-eafd-account-create-h97jm"] Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.378015 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.394584 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-cf80-account-create-zz7cn"] Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.394656 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96dcm\" (UniqueName: \"kubernetes.io/projected/bcadc39c-6259-49a9-9f5f-38a80545b5a4-kube-api-access-96dcm\") pod \"barbican-eafd-account-create-h97jm\" (UID: \"bcadc39c-6259-49a9-9f5f-38a80545b5a4\") " pod="openstack/barbican-eafd-account-create-h97jm" Oct 01 09:52:08 crc kubenswrapper[4787]: E1001 09:52:08.395052 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55091ec1-5a26-4825-9a61-a49cecda5a6f" containerName="glance-log" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.395095 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="55091ec1-5a26-4825-9a61-a49cecda5a6f" containerName="glance-log" Oct 01 09:52:08 crc kubenswrapper[4787]: E1001 09:52:08.395137 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55091ec1-5a26-4825-9a61-a49cecda5a6f" containerName="glance-httpd" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.395146 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="55091ec1-5a26-4825-9a61-a49cecda5a6f" containerName="glance-httpd" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.395361 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="55091ec1-5a26-4825-9a61-a49cecda5a6f" containerName="glance-httpd" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.395393 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="55091ec1-5a26-4825-9a61-a49cecda5a6f" containerName="glance-log" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.396666 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-cf80-account-create-zz7cn" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.399189 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.405014 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-cf80-account-create-zz7cn"] Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.496000 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55091ec1-5a26-4825-9a61-a49cecda5a6f-scripts\") pod \"55091ec1-5a26-4825-9a61-a49cecda5a6f\" (UID: \"55091ec1-5a26-4825-9a61-a49cecda5a6f\") " Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.496210 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/55091ec1-5a26-4825-9a61-a49cecda5a6f-logs\") pod \"55091ec1-5a26-4825-9a61-a49cecda5a6f\" (UID: \"55091ec1-5a26-4825-9a61-a49cecda5a6f\") " Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.496278 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55091ec1-5a26-4825-9a61-a49cecda5a6f-combined-ca-bundle\") pod \"55091ec1-5a26-4825-9a61-a49cecda5a6f\" (UID: \"55091ec1-5a26-4825-9a61-a49cecda5a6f\") " Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.496459 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/55091ec1-5a26-4825-9a61-a49cecda5a6f-httpd-run\") pod \"55091ec1-5a26-4825-9a61-a49cecda5a6f\" (UID: \"55091ec1-5a26-4825-9a61-a49cecda5a6f\") " Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.496483 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"55091ec1-5a26-4825-9a61-a49cecda5a6f\" (UID: \"55091ec1-5a26-4825-9a61-a49cecda5a6f\") " Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.496540 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55091ec1-5a26-4825-9a61-a49cecda5a6f-config-data\") pod \"55091ec1-5a26-4825-9a61-a49cecda5a6f\" (UID: \"55091ec1-5a26-4825-9a61-a49cecda5a6f\") " Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.496568 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qllv9\" (UniqueName: \"kubernetes.io/projected/55091ec1-5a26-4825-9a61-a49cecda5a6f-kube-api-access-qllv9\") pod \"55091ec1-5a26-4825-9a61-a49cecda5a6f\" (UID: \"55091ec1-5a26-4825-9a61-a49cecda5a6f\") " Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.502098 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjbdh\" (UniqueName: \"kubernetes.io/projected/54517204-c321-4959-851b-9c642f07a7a7-kube-api-access-zjbdh\") pod \"cinder-cf80-account-create-zz7cn\" (UID: \"54517204-c321-4959-851b-9c642f07a7a7\") " pod="openstack/cinder-cf80-account-create-zz7cn" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.502320 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55091ec1-5a26-4825-9a61-a49cecda5a6f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "55091ec1-5a26-4825-9a61-a49cecda5a6f" (UID: "55091ec1-5a26-4825-9a61-a49cecda5a6f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.502537 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96dcm\" (UniqueName: \"kubernetes.io/projected/bcadc39c-6259-49a9-9f5f-38a80545b5a4-kube-api-access-96dcm\") pod \"barbican-eafd-account-create-h97jm\" (UID: \"bcadc39c-6259-49a9-9f5f-38a80545b5a4\") " pod="openstack/barbican-eafd-account-create-h97jm" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.502883 4787 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/55091ec1-5a26-4825-9a61-a49cecda5a6f-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.518143 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55091ec1-5a26-4825-9a61-a49cecda5a6f-kube-api-access-qllv9" (OuterVolumeSpecName: "kube-api-access-qllv9") pod "55091ec1-5a26-4825-9a61-a49cecda5a6f" (UID: "55091ec1-5a26-4825-9a61-a49cecda5a6f"). InnerVolumeSpecName "kube-api-access-qllv9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.529318 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55091ec1-5a26-4825-9a61-a49cecda5a6f-logs" (OuterVolumeSpecName: "logs") pod "55091ec1-5a26-4825-9a61-a49cecda5a6f" (UID: "55091ec1-5a26-4825-9a61-a49cecda5a6f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.535144 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "55091ec1-5a26-4825-9a61-a49cecda5a6f" (UID: "55091ec1-5a26-4825-9a61-a49cecda5a6f"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.542650 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96dcm\" (UniqueName: \"kubernetes.io/projected/bcadc39c-6259-49a9-9f5f-38a80545b5a4-kube-api-access-96dcm\") pod \"barbican-eafd-account-create-h97jm\" (UID: \"bcadc39c-6259-49a9-9f5f-38a80545b5a4\") " pod="openstack/barbican-eafd-account-create-h97jm" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.556935 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55091ec1-5a26-4825-9a61-a49cecda5a6f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "55091ec1-5a26-4825-9a61-a49cecda5a6f" (UID: "55091ec1-5a26-4825-9a61-a49cecda5a6f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.562220 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55091ec1-5a26-4825-9a61-a49cecda5a6f-scripts" (OuterVolumeSpecName: "scripts") pod "55091ec1-5a26-4825-9a61-a49cecda5a6f" (UID: "55091ec1-5a26-4825-9a61-a49cecda5a6f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.586951 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55091ec1-5a26-4825-9a61-a49cecda5a6f-config-data" (OuterVolumeSpecName: "config-data") pod "55091ec1-5a26-4825-9a61-a49cecda5a6f" (UID: "55091ec1-5a26-4825-9a61-a49cecda5a6f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.604637 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjbdh\" (UniqueName: \"kubernetes.io/projected/54517204-c321-4959-851b-9c642f07a7a7-kube-api-access-zjbdh\") pod \"cinder-cf80-account-create-zz7cn\" (UID: \"54517204-c321-4959-851b-9c642f07a7a7\") " pod="openstack/cinder-cf80-account-create-zz7cn" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.604778 4787 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/55091ec1-5a26-4825-9a61-a49cecda5a6f-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.604794 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55091ec1-5a26-4825-9a61-a49cecda5a6f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.604819 4787 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.604830 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55091ec1-5a26-4825-9a61-a49cecda5a6f-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.604840 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qllv9\" (UniqueName: \"kubernetes.io/projected/55091ec1-5a26-4825-9a61-a49cecda5a6f-kube-api-access-qllv9\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.604849 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55091ec1-5a26-4825-9a61-a49cecda5a6f-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.625747 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjbdh\" (UniqueName: \"kubernetes.io/projected/54517204-c321-4959-851b-9c642f07a7a7-kube-api-access-zjbdh\") pod \"cinder-cf80-account-create-zz7cn\" (UID: \"54517204-c321-4959-851b-9c642f07a7a7\") " pod="openstack/cinder-cf80-account-create-zz7cn" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.627909 4787 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.697680 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-eafd-account-create-h97jm" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.703744 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-819c-account-create-86wn2"] Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.705437 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-819c-account-create-86wn2" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.708181 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.710538 4787 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.714395 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-819c-account-create-86wn2"] Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.715176 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-cf80-account-create-zz7cn" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.812325 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dmmj\" (UniqueName: \"kubernetes.io/projected/22964b11-e82b-4dfa-ac42-70005cbb5caf-kube-api-access-2dmmj\") pod \"neutron-819c-account-create-86wn2\" (UID: \"22964b11-e82b-4dfa-ac42-70005cbb5caf\") " pod="openstack/neutron-819c-account-create-86wn2" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.914962 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dmmj\" (UniqueName: \"kubernetes.io/projected/22964b11-e82b-4dfa-ac42-70005cbb5caf-kube-api-access-2dmmj\") pod \"neutron-819c-account-create-86wn2\" (UID: \"22964b11-e82b-4dfa-ac42-70005cbb5caf\") " pod="openstack/neutron-819c-account-create-86wn2" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.948488 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dmmj\" (UniqueName: \"kubernetes.io/projected/22964b11-e82b-4dfa-ac42-70005cbb5caf-kube-api-access-2dmmj\") pod \"neutron-819c-account-create-86wn2\" (UID: \"22964b11-e82b-4dfa-ac42-70005cbb5caf\") " pod="openstack/neutron-819c-account-create-86wn2" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.968357 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.968408 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"55091ec1-5a26-4825-9a61-a49cecda5a6f","Type":"ContainerDied","Data":"ee2f77af998f7916b9be643dcc9322d0ea37a6c82778d6596e66bb6ad73b5c7f"} Oct 01 09:52:08 crc kubenswrapper[4787]: I1001 09:52:08.968462 4787 scope.go:117] "RemoveContainer" containerID="0256ce9b0aaf2dc94f1c6ed17241671d2db6b17016086397971f59f0401d3e93" Oct 01 09:52:09 crc kubenswrapper[4787]: I1001 09:52:09.013022 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:52:09 crc kubenswrapper[4787]: I1001 09:52:09.024800 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-819c-account-create-86wn2" Oct 01 09:52:09 crc kubenswrapper[4787]: I1001 09:52:09.032780 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:52:09 crc kubenswrapper[4787]: I1001 09:52:09.046182 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:52:09 crc kubenswrapper[4787]: I1001 09:52:09.047959 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 09:52:09 crc kubenswrapper[4787]: I1001 09:52:09.050215 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 01 09:52:09 crc kubenswrapper[4787]: I1001 09:52:09.055783 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:52:09 crc kubenswrapper[4787]: I1001 09:52:09.120858 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/556ee9b9-8f10-4de4-9a97-1d3327f885f1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"556ee9b9-8f10-4de4-9a97-1d3327f885f1\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:09 crc kubenswrapper[4787]: I1001 09:52:09.121333 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/556ee9b9-8f10-4de4-9a97-1d3327f885f1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"556ee9b9-8f10-4de4-9a97-1d3327f885f1\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:09 crc kubenswrapper[4787]: I1001 09:52:09.121357 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/556ee9b9-8f10-4de4-9a97-1d3327f885f1-logs\") pod \"glance-default-internal-api-0\" (UID: \"556ee9b9-8f10-4de4-9a97-1d3327f885f1\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:09 crc kubenswrapper[4787]: I1001 09:52:09.121401 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"556ee9b9-8f10-4de4-9a97-1d3327f885f1\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:09 crc kubenswrapper[4787]: I1001 09:52:09.121474 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/556ee9b9-8f10-4de4-9a97-1d3327f885f1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"556ee9b9-8f10-4de4-9a97-1d3327f885f1\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:09 crc kubenswrapper[4787]: I1001 09:52:09.121505 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/556ee9b9-8f10-4de4-9a97-1d3327f885f1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"556ee9b9-8f10-4de4-9a97-1d3327f885f1\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:09 crc kubenswrapper[4787]: I1001 09:52:09.121535 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwvtw\" (UniqueName: \"kubernetes.io/projected/556ee9b9-8f10-4de4-9a97-1d3327f885f1-kube-api-access-hwvtw\") pod \"glance-default-internal-api-0\" (UID: \"556ee9b9-8f10-4de4-9a97-1d3327f885f1\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:09 crc kubenswrapper[4787]: I1001 09:52:09.228297 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/556ee9b9-8f10-4de4-9a97-1d3327f885f1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"556ee9b9-8f10-4de4-9a97-1d3327f885f1\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:09 crc kubenswrapper[4787]: I1001 09:52:09.228366 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/556ee9b9-8f10-4de4-9a97-1d3327f885f1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"556ee9b9-8f10-4de4-9a97-1d3327f885f1\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:09 crc kubenswrapper[4787]: I1001 09:52:09.228389 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/556ee9b9-8f10-4de4-9a97-1d3327f885f1-logs\") pod \"glance-default-internal-api-0\" (UID: \"556ee9b9-8f10-4de4-9a97-1d3327f885f1\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:09 crc kubenswrapper[4787]: I1001 09:52:09.228433 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"556ee9b9-8f10-4de4-9a97-1d3327f885f1\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:09 crc kubenswrapper[4787]: I1001 09:52:09.228483 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/556ee9b9-8f10-4de4-9a97-1d3327f885f1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"556ee9b9-8f10-4de4-9a97-1d3327f885f1\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:09 crc kubenswrapper[4787]: I1001 09:52:09.228506 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/556ee9b9-8f10-4de4-9a97-1d3327f885f1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"556ee9b9-8f10-4de4-9a97-1d3327f885f1\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:09 crc kubenswrapper[4787]: I1001 09:52:09.228533 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwvtw\" (UniqueName: \"kubernetes.io/projected/556ee9b9-8f10-4de4-9a97-1d3327f885f1-kube-api-access-hwvtw\") pod \"glance-default-internal-api-0\" (UID: \"556ee9b9-8f10-4de4-9a97-1d3327f885f1\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:09 crc kubenswrapper[4787]: I1001 09:52:09.228918 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"556ee9b9-8f10-4de4-9a97-1d3327f885f1\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Oct 01 09:52:09 crc kubenswrapper[4787]: I1001 09:52:09.229569 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/556ee9b9-8f10-4de4-9a97-1d3327f885f1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"556ee9b9-8f10-4de4-9a97-1d3327f885f1\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:09 crc kubenswrapper[4787]: I1001 09:52:09.232639 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/556ee9b9-8f10-4de4-9a97-1d3327f885f1-logs\") pod \"glance-default-internal-api-0\" (UID: \"556ee9b9-8f10-4de4-9a97-1d3327f885f1\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:09 crc kubenswrapper[4787]: I1001 09:52:09.238428 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/556ee9b9-8f10-4de4-9a97-1d3327f885f1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"556ee9b9-8f10-4de4-9a97-1d3327f885f1\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:09 crc kubenswrapper[4787]: I1001 09:52:09.239298 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/556ee9b9-8f10-4de4-9a97-1d3327f885f1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"556ee9b9-8f10-4de4-9a97-1d3327f885f1\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:09 crc kubenswrapper[4787]: I1001 09:52:09.240953 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/556ee9b9-8f10-4de4-9a97-1d3327f885f1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"556ee9b9-8f10-4de4-9a97-1d3327f885f1\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:09 crc kubenswrapper[4787]: I1001 09:52:09.250490 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwvtw\" (UniqueName: \"kubernetes.io/projected/556ee9b9-8f10-4de4-9a97-1d3327f885f1-kube-api-access-hwvtw\") pod \"glance-default-internal-api-0\" (UID: \"556ee9b9-8f10-4de4-9a97-1d3327f885f1\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:09 crc kubenswrapper[4787]: I1001 09:52:09.255352 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"556ee9b9-8f10-4de4-9a97-1d3327f885f1\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:09 crc kubenswrapper[4787]: I1001 09:52:09.366740 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 09:52:10 crc kubenswrapper[4787]: I1001 09:52:10.540630 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55091ec1-5a26-4825-9a61-a49cecda5a6f" path="/var/lib/kubelet/pods/55091ec1-5a26-4825-9a61-a49cecda5a6f/volumes" Oct 01 09:52:11 crc kubenswrapper[4787]: I1001 09:52:11.697254 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7bb97fccb5-vblx7" Oct 01 09:52:11 crc kubenswrapper[4787]: I1001 09:52:11.811297 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c6d5d5bd7-4mbvj"] Oct 01 09:52:11 crc kubenswrapper[4787]: I1001 09:52:11.811747 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6c6d5d5bd7-4mbvj" podUID="bbade729-c3cb-447b-bdd7-251fbc80c454" containerName="dnsmasq-dns" containerID="cri-o://6020b30fd47f95aa07a6b93dfa13052657b3120e459f6fe4674a14f5b6486a9c" gracePeriod=10 Oct 01 09:52:11 crc kubenswrapper[4787]: I1001 09:52:11.877854 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:52:11 crc kubenswrapper[4787]: I1001 09:52:11.999364 4787 generic.go:334] "Generic (PLEG): container finished" podID="bbade729-c3cb-447b-bdd7-251fbc80c454" containerID="6020b30fd47f95aa07a6b93dfa13052657b3120e459f6fe4674a14f5b6486a9c" exitCode=0 Oct 01 09:52:11 crc kubenswrapper[4787]: I1001 09:52:11.999413 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c6d5d5bd7-4mbvj" event={"ID":"bbade729-c3cb-447b-bdd7-251fbc80c454","Type":"ContainerDied","Data":"6020b30fd47f95aa07a6b93dfa13052657b3120e459f6fe4674a14f5b6486a9c"} Oct 01 09:52:12 crc kubenswrapper[4787]: E1001 09:52:12.110799 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbbade729_c3cb_447b_bdd7_251fbc80c454.slice/crio-conmon-6020b30fd47f95aa07a6b93dfa13052657b3120e459f6fe4674a14f5b6486a9c.scope\": RecentStats: unable to find data in memory cache]" Oct 01 09:52:12 crc kubenswrapper[4787]: I1001 09:52:12.733844 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-b48666669-4sqql"] Oct 01 09:52:12 crc kubenswrapper[4787]: I1001 09:52:12.801650 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-67879968b-d5598"] Oct 01 09:52:12 crc kubenswrapper[4787]: I1001 09:52:12.813462 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67879968b-d5598" Oct 01 09:52:12 crc kubenswrapper[4787]: I1001 09:52:12.819461 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Oct 01 09:52:12 crc kubenswrapper[4787]: I1001 09:52:12.833001 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-67879968b-d5598"] Oct 01 09:52:12 crc kubenswrapper[4787]: I1001 09:52:12.905834 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gxpk\" (UniqueName: \"kubernetes.io/projected/7983e213-c317-42b2-8944-85bdb78dac8f-kube-api-access-5gxpk\") pod \"horizon-67879968b-d5598\" (UID: \"7983e213-c317-42b2-8944-85bdb78dac8f\") " pod="openstack/horizon-67879968b-d5598" Oct 01 09:52:12 crc kubenswrapper[4787]: I1001 09:52:12.905996 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7983e213-c317-42b2-8944-85bdb78dac8f-horizon-secret-key\") pod \"horizon-67879968b-d5598\" (UID: \"7983e213-c317-42b2-8944-85bdb78dac8f\") " pod="openstack/horizon-67879968b-d5598" Oct 01 09:52:12 crc kubenswrapper[4787]: I1001 09:52:12.906222 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7983e213-c317-42b2-8944-85bdb78dac8f-scripts\") pod \"horizon-67879968b-d5598\" (UID: \"7983e213-c317-42b2-8944-85bdb78dac8f\") " pod="openstack/horizon-67879968b-d5598" Oct 01 09:52:12 crc kubenswrapper[4787]: I1001 09:52:12.906466 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7983e213-c317-42b2-8944-85bdb78dac8f-logs\") pod \"horizon-67879968b-d5598\" (UID: \"7983e213-c317-42b2-8944-85bdb78dac8f\") " pod="openstack/horizon-67879968b-d5598" Oct 01 09:52:12 crc kubenswrapper[4787]: I1001 09:52:12.906696 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/7983e213-c317-42b2-8944-85bdb78dac8f-horizon-tls-certs\") pod \"horizon-67879968b-d5598\" (UID: \"7983e213-c317-42b2-8944-85bdb78dac8f\") " pod="openstack/horizon-67879968b-d5598" Oct 01 09:52:12 crc kubenswrapper[4787]: I1001 09:52:12.910527 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7983e213-c317-42b2-8944-85bdb78dac8f-combined-ca-bundle\") pod \"horizon-67879968b-d5598\" (UID: \"7983e213-c317-42b2-8944-85bdb78dac8f\") " pod="openstack/horizon-67879968b-d5598" Oct 01 09:52:12 crc kubenswrapper[4787]: I1001 09:52:12.910695 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7983e213-c317-42b2-8944-85bdb78dac8f-config-data\") pod \"horizon-67879968b-d5598\" (UID: \"7983e213-c317-42b2-8944-85bdb78dac8f\") " pod="openstack/horizon-67879968b-d5598" Oct 01 09:52:12 crc kubenswrapper[4787]: I1001 09:52:12.912966 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7565dcfdb5-mcrm8"] Oct 01 09:52:12 crc kubenswrapper[4787]: I1001 09:52:12.928112 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7d65d5957b-44rtv"] Oct 01 09:52:12 crc kubenswrapper[4787]: I1001 09:52:12.929690 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d65d5957b-44rtv" Oct 01 09:52:12 crc kubenswrapper[4787]: I1001 09:52:12.951212 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7d65d5957b-44rtv"] Oct 01 09:52:13 crc kubenswrapper[4787]: I1001 09:52:13.012121 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7983e213-c317-42b2-8944-85bdb78dac8f-combined-ca-bundle\") pod \"horizon-67879968b-d5598\" (UID: \"7983e213-c317-42b2-8944-85bdb78dac8f\") " pod="openstack/horizon-67879968b-d5598" Oct 01 09:52:13 crc kubenswrapper[4787]: I1001 09:52:13.012177 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/db1d14ba-8d00-4508-9c89-fdadb562ad0d-horizon-tls-certs\") pod \"horizon-7d65d5957b-44rtv\" (UID: \"db1d14ba-8d00-4508-9c89-fdadb562ad0d\") " pod="openstack/horizon-7d65d5957b-44rtv" Oct 01 09:52:13 crc kubenswrapper[4787]: I1001 09:52:13.012209 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7983e213-c317-42b2-8944-85bdb78dac8f-config-data\") pod \"horizon-67879968b-d5598\" (UID: \"7983e213-c317-42b2-8944-85bdb78dac8f\") " pod="openstack/horizon-67879968b-d5598" Oct 01 09:52:13 crc kubenswrapper[4787]: I1001 09:52:13.012230 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/db1d14ba-8d00-4508-9c89-fdadb562ad0d-scripts\") pod \"horizon-7d65d5957b-44rtv\" (UID: \"db1d14ba-8d00-4508-9c89-fdadb562ad0d\") " pod="openstack/horizon-7d65d5957b-44rtv" Oct 01 09:52:13 crc kubenswrapper[4787]: I1001 09:52:13.012253 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5fpz\" (UniqueName: \"kubernetes.io/projected/db1d14ba-8d00-4508-9c89-fdadb562ad0d-kube-api-access-d5fpz\") pod \"horizon-7d65d5957b-44rtv\" (UID: \"db1d14ba-8d00-4508-9c89-fdadb562ad0d\") " pod="openstack/horizon-7d65d5957b-44rtv" Oct 01 09:52:13 crc kubenswrapper[4787]: I1001 09:52:13.012279 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gxpk\" (UniqueName: \"kubernetes.io/projected/7983e213-c317-42b2-8944-85bdb78dac8f-kube-api-access-5gxpk\") pod \"horizon-67879968b-d5598\" (UID: \"7983e213-c317-42b2-8944-85bdb78dac8f\") " pod="openstack/horizon-67879968b-d5598" Oct 01 09:52:13 crc kubenswrapper[4787]: I1001 09:52:13.012302 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7983e213-c317-42b2-8944-85bdb78dac8f-horizon-secret-key\") pod \"horizon-67879968b-d5598\" (UID: \"7983e213-c317-42b2-8944-85bdb78dac8f\") " pod="openstack/horizon-67879968b-d5598" Oct 01 09:52:13 crc kubenswrapper[4787]: I1001 09:52:13.012320 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/db1d14ba-8d00-4508-9c89-fdadb562ad0d-horizon-secret-key\") pod \"horizon-7d65d5957b-44rtv\" (UID: \"db1d14ba-8d00-4508-9c89-fdadb562ad0d\") " pod="openstack/horizon-7d65d5957b-44rtv" Oct 01 09:52:13 crc kubenswrapper[4787]: I1001 09:52:13.012339 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db1d14ba-8d00-4508-9c89-fdadb562ad0d-logs\") pod \"horizon-7d65d5957b-44rtv\" (UID: \"db1d14ba-8d00-4508-9c89-fdadb562ad0d\") " pod="openstack/horizon-7d65d5957b-44rtv" Oct 01 09:52:13 crc kubenswrapper[4787]: I1001 09:52:13.012358 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7983e213-c317-42b2-8944-85bdb78dac8f-scripts\") pod \"horizon-67879968b-d5598\" (UID: \"7983e213-c317-42b2-8944-85bdb78dac8f\") " pod="openstack/horizon-67879968b-d5598" Oct 01 09:52:13 crc kubenswrapper[4787]: I1001 09:52:13.012397 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7983e213-c317-42b2-8944-85bdb78dac8f-logs\") pod \"horizon-67879968b-d5598\" (UID: \"7983e213-c317-42b2-8944-85bdb78dac8f\") " pod="openstack/horizon-67879968b-d5598" Oct 01 09:52:13 crc kubenswrapper[4787]: I1001 09:52:13.012418 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db1d14ba-8d00-4508-9c89-fdadb562ad0d-combined-ca-bundle\") pod \"horizon-7d65d5957b-44rtv\" (UID: \"db1d14ba-8d00-4508-9c89-fdadb562ad0d\") " pod="openstack/horizon-7d65d5957b-44rtv" Oct 01 09:52:13 crc kubenswrapper[4787]: I1001 09:52:13.012450 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/db1d14ba-8d00-4508-9c89-fdadb562ad0d-config-data\") pod \"horizon-7d65d5957b-44rtv\" (UID: \"db1d14ba-8d00-4508-9c89-fdadb562ad0d\") " pod="openstack/horizon-7d65d5957b-44rtv" Oct 01 09:52:13 crc kubenswrapper[4787]: I1001 09:52:13.012467 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/7983e213-c317-42b2-8944-85bdb78dac8f-horizon-tls-certs\") pod \"horizon-67879968b-d5598\" (UID: \"7983e213-c317-42b2-8944-85bdb78dac8f\") " pod="openstack/horizon-67879968b-d5598" Oct 01 09:52:13 crc kubenswrapper[4787]: I1001 09:52:13.014165 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7983e213-c317-42b2-8944-85bdb78dac8f-logs\") pod \"horizon-67879968b-d5598\" (UID: \"7983e213-c317-42b2-8944-85bdb78dac8f\") " pod="openstack/horizon-67879968b-d5598" Oct 01 09:52:13 crc kubenswrapper[4787]: I1001 09:52:13.015619 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7983e213-c317-42b2-8944-85bdb78dac8f-scripts\") pod \"horizon-67879968b-d5598\" (UID: \"7983e213-c317-42b2-8944-85bdb78dac8f\") " pod="openstack/horizon-67879968b-d5598" Oct 01 09:52:13 crc kubenswrapper[4787]: I1001 09:52:13.015674 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7983e213-c317-42b2-8944-85bdb78dac8f-config-data\") pod \"horizon-67879968b-d5598\" (UID: \"7983e213-c317-42b2-8944-85bdb78dac8f\") " pod="openstack/horizon-67879968b-d5598" Oct 01 09:52:13 crc kubenswrapper[4787]: I1001 09:52:13.024960 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/7983e213-c317-42b2-8944-85bdb78dac8f-horizon-tls-certs\") pod \"horizon-67879968b-d5598\" (UID: \"7983e213-c317-42b2-8944-85bdb78dac8f\") " pod="openstack/horizon-67879968b-d5598" Oct 01 09:52:13 crc kubenswrapper[4787]: I1001 09:52:13.034469 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7983e213-c317-42b2-8944-85bdb78dac8f-horizon-secret-key\") pod \"horizon-67879968b-d5598\" (UID: \"7983e213-c317-42b2-8944-85bdb78dac8f\") " pod="openstack/horizon-67879968b-d5598" Oct 01 09:52:13 crc kubenswrapper[4787]: I1001 09:52:13.043664 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gxpk\" (UniqueName: \"kubernetes.io/projected/7983e213-c317-42b2-8944-85bdb78dac8f-kube-api-access-5gxpk\") pod \"horizon-67879968b-d5598\" (UID: \"7983e213-c317-42b2-8944-85bdb78dac8f\") " pod="openstack/horizon-67879968b-d5598" Oct 01 09:52:13 crc kubenswrapper[4787]: I1001 09:52:13.064843 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7983e213-c317-42b2-8944-85bdb78dac8f-combined-ca-bundle\") pod \"horizon-67879968b-d5598\" (UID: \"7983e213-c317-42b2-8944-85bdb78dac8f\") " pod="openstack/horizon-67879968b-d5598" Oct 01 09:52:13 crc kubenswrapper[4787]: I1001 09:52:13.114244 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/db1d14ba-8d00-4508-9c89-fdadb562ad0d-horizon-tls-certs\") pod \"horizon-7d65d5957b-44rtv\" (UID: \"db1d14ba-8d00-4508-9c89-fdadb562ad0d\") " pod="openstack/horizon-7d65d5957b-44rtv" Oct 01 09:52:13 crc kubenswrapper[4787]: I1001 09:52:13.114311 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/db1d14ba-8d00-4508-9c89-fdadb562ad0d-scripts\") pod \"horizon-7d65d5957b-44rtv\" (UID: \"db1d14ba-8d00-4508-9c89-fdadb562ad0d\") " pod="openstack/horizon-7d65d5957b-44rtv" Oct 01 09:52:13 crc kubenswrapper[4787]: I1001 09:52:13.114342 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5fpz\" (UniqueName: \"kubernetes.io/projected/db1d14ba-8d00-4508-9c89-fdadb562ad0d-kube-api-access-d5fpz\") pod \"horizon-7d65d5957b-44rtv\" (UID: \"db1d14ba-8d00-4508-9c89-fdadb562ad0d\") " pod="openstack/horizon-7d65d5957b-44rtv" Oct 01 09:52:13 crc kubenswrapper[4787]: I1001 09:52:13.114372 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/db1d14ba-8d00-4508-9c89-fdadb562ad0d-horizon-secret-key\") pod \"horizon-7d65d5957b-44rtv\" (UID: \"db1d14ba-8d00-4508-9c89-fdadb562ad0d\") " pod="openstack/horizon-7d65d5957b-44rtv" Oct 01 09:52:13 crc kubenswrapper[4787]: I1001 09:52:13.114390 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db1d14ba-8d00-4508-9c89-fdadb562ad0d-logs\") pod \"horizon-7d65d5957b-44rtv\" (UID: \"db1d14ba-8d00-4508-9c89-fdadb562ad0d\") " pod="openstack/horizon-7d65d5957b-44rtv" Oct 01 09:52:13 crc kubenswrapper[4787]: I1001 09:52:13.114439 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db1d14ba-8d00-4508-9c89-fdadb562ad0d-combined-ca-bundle\") pod \"horizon-7d65d5957b-44rtv\" (UID: \"db1d14ba-8d00-4508-9c89-fdadb562ad0d\") " pod="openstack/horizon-7d65d5957b-44rtv" Oct 01 09:52:13 crc kubenswrapper[4787]: I1001 09:52:13.114471 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/db1d14ba-8d00-4508-9c89-fdadb562ad0d-config-data\") pod \"horizon-7d65d5957b-44rtv\" (UID: \"db1d14ba-8d00-4508-9c89-fdadb562ad0d\") " pod="openstack/horizon-7d65d5957b-44rtv" Oct 01 09:52:13 crc kubenswrapper[4787]: I1001 09:52:13.144904 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/db1d14ba-8d00-4508-9c89-fdadb562ad0d-horizon-tls-certs\") pod \"horizon-7d65d5957b-44rtv\" (UID: \"db1d14ba-8d00-4508-9c89-fdadb562ad0d\") " pod="openstack/horizon-7d65d5957b-44rtv" Oct 01 09:52:13 crc kubenswrapper[4787]: I1001 09:52:13.145354 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/db1d14ba-8d00-4508-9c89-fdadb562ad0d-config-data\") pod \"horizon-7d65d5957b-44rtv\" (UID: \"db1d14ba-8d00-4508-9c89-fdadb562ad0d\") " pod="openstack/horizon-7d65d5957b-44rtv" Oct 01 09:52:13 crc kubenswrapper[4787]: I1001 09:52:13.145562 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/db1d14ba-8d00-4508-9c89-fdadb562ad0d-horizon-secret-key\") pod \"horizon-7d65d5957b-44rtv\" (UID: \"db1d14ba-8d00-4508-9c89-fdadb562ad0d\") " pod="openstack/horizon-7d65d5957b-44rtv" Oct 01 09:52:13 crc kubenswrapper[4787]: I1001 09:52:13.145753 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/db1d14ba-8d00-4508-9c89-fdadb562ad0d-scripts\") pod \"horizon-7d65d5957b-44rtv\" (UID: \"db1d14ba-8d00-4508-9c89-fdadb562ad0d\") " pod="openstack/horizon-7d65d5957b-44rtv" Oct 01 09:52:13 crc kubenswrapper[4787]: I1001 09:52:13.145767 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db1d14ba-8d00-4508-9c89-fdadb562ad0d-logs\") pod \"horizon-7d65d5957b-44rtv\" (UID: \"db1d14ba-8d00-4508-9c89-fdadb562ad0d\") " pod="openstack/horizon-7d65d5957b-44rtv" Oct 01 09:52:13 crc kubenswrapper[4787]: I1001 09:52:13.146608 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db1d14ba-8d00-4508-9c89-fdadb562ad0d-combined-ca-bundle\") pod \"horizon-7d65d5957b-44rtv\" (UID: \"db1d14ba-8d00-4508-9c89-fdadb562ad0d\") " pod="openstack/horizon-7d65d5957b-44rtv" Oct 01 09:52:13 crc kubenswrapper[4787]: I1001 09:52:13.149284 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5fpz\" (UniqueName: \"kubernetes.io/projected/db1d14ba-8d00-4508-9c89-fdadb562ad0d-kube-api-access-d5fpz\") pod \"horizon-7d65d5957b-44rtv\" (UID: \"db1d14ba-8d00-4508-9c89-fdadb562ad0d\") " pod="openstack/horizon-7d65d5957b-44rtv" Oct 01 09:52:13 crc kubenswrapper[4787]: I1001 09:52:13.174595 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67879968b-d5598" Oct 01 09:52:13 crc kubenswrapper[4787]: I1001 09:52:13.278337 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d65d5957b-44rtv" Oct 01 09:52:13 crc kubenswrapper[4787]: I1001 09:52:13.755735 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6c6d5d5bd7-4mbvj" podUID="bbade729-c3cb-447b-bdd7-251fbc80c454" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.116:5353: connect: connection refused" Oct 01 09:52:18 crc kubenswrapper[4787]: I1001 09:52:18.755685 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6c6d5d5bd7-4mbvj" podUID="bbade729-c3cb-447b-bdd7-251fbc80c454" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.116:5353: connect: connection refused" Oct 01 09:52:21 crc kubenswrapper[4787]: E1001 09:52:21.988461 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon@sha256:9470db6caf5102cf37ddb1f137f17b05ef7119f174f4189beb4839ef7f65730c" Oct 01 09:52:21 crc kubenswrapper[4787]: E1001 09:52:21.989020 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon@sha256:9470db6caf5102cf37ddb1f137f17b05ef7119f174f4189beb4839ef7f65730c,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n89h68bhb5h668h5fbh68ch556h6fh57bh666h98h577hd4h556h54dh8bh575h5f4hddh649h55fh5ch66fh58dh7ch97h7h5ffh74h585h9dh9cq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gvxbh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-b48666669-4sqql_openstack(3869928b-1191-433d-87bf-7187615b57e0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 09:52:21 crc kubenswrapper[4787]: E1001 09:52:21.991954 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon@sha256:9470db6caf5102cf37ddb1f137f17b05ef7119f174f4189beb4839ef7f65730c" Oct 01 09:52:21 crc kubenswrapper[4787]: E1001 09:52:21.992316 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon@sha256:9470db6caf5102cf37ddb1f137f17b05ef7119f174f4189beb4839ef7f65730c\\\"\"]" pod="openstack/horizon-b48666669-4sqql" podUID="3869928b-1191-433d-87bf-7187615b57e0" Oct 01 09:52:21 crc kubenswrapper[4787]: E1001 09:52:21.992392 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon@sha256:9470db6caf5102cf37ddb1f137f17b05ef7119f174f4189beb4839ef7f65730c,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nf7h54dh5d4h669h5b4h6bh58bhcfh66h666h66chc4h5d6h59dh7ch59h6dh569hf4h597h584hd8h5dch57bh87h54ch6fhcfh5dch99h97h68q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mdpd2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-7565dcfdb5-mcrm8_openstack(54c87b91-944a-4f8c-a556-21649f3b9b1b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 09:52:21 crc kubenswrapper[4787]: E1001 09:52:21.994449 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon@sha256:9470db6caf5102cf37ddb1f137f17b05ef7119f174f4189beb4839ef7f65730c\\\"\"]" pod="openstack/horizon-7565dcfdb5-mcrm8" podUID="54c87b91-944a-4f8c-a556-21649f3b9b1b" Oct 01 09:52:22 crc kubenswrapper[4787]: E1001 09:52:22.011121 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon@sha256:9470db6caf5102cf37ddb1f137f17b05ef7119f174f4189beb4839ef7f65730c" Oct 01 09:52:22 crc kubenswrapper[4787]: E1001 09:52:22.011328 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon@sha256:9470db6caf5102cf37ddb1f137f17b05ef7119f174f4189beb4839ef7f65730c,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n96h65dh699h89h55h569h646h56dh658h5dhc8h67h5b5hfh9ch58ch5fch5bdh5d5h7fh565h654h55chf7h67bh54hch686h675h58ch588h68cq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vncx9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-58779b6b6f-rqbn7_openstack(2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 09:52:22 crc kubenswrapper[4787]: E1001 09:52:22.014032 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon@sha256:9470db6caf5102cf37ddb1f137f17b05ef7119f174f4189beb4839ef7f65730c\\\"\"]" pod="openstack/horizon-58779b6b6f-rqbn7" podUID="2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.084748 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.092382 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-tk8w8" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.134101 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.134061 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3","Type":"ContainerDied","Data":"9900a4add31c368beea14cfcefbd6807b21c8275eb21edd9ea8ae4f00d6d8964"} Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.137131 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-tk8w8" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.137172 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-tk8w8" event={"ID":"266f85d9-996e-41a9-a83d-5451e74689b6","Type":"ContainerDied","Data":"42049a54e6439475e05e6c3b439358d6220f361271336081ec8f5793e3599625"} Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.137243 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42049a54e6439475e05e6c3b439358d6220f361271336081ec8f5793e3599625" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.198043 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/266f85d9-996e-41a9-a83d-5451e74689b6-credential-keys\") pod \"266f85d9-996e-41a9-a83d-5451e74689b6\" (UID: \"266f85d9-996e-41a9-a83d-5451e74689b6\") " Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.198149 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/266f85d9-996e-41a9-a83d-5451e74689b6-config-data\") pod \"266f85d9-996e-41a9-a83d-5451e74689b6\" (UID: \"266f85d9-996e-41a9-a83d-5451e74689b6\") " Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.198181 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3-scripts\") pod \"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3\" (UID: \"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3\") " Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.198205 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/266f85d9-996e-41a9-a83d-5451e74689b6-combined-ca-bundle\") pod \"266f85d9-996e-41a9-a83d-5451e74689b6\" (UID: \"266f85d9-996e-41a9-a83d-5451e74689b6\") " Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.198234 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3-logs\") pod \"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3\" (UID: \"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3\") " Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.198264 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3\" (UID: \"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3\") " Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.198325 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jfkhh\" (UniqueName: \"kubernetes.io/projected/266f85d9-996e-41a9-a83d-5451e74689b6-kube-api-access-jfkhh\") pod \"266f85d9-996e-41a9-a83d-5451e74689b6\" (UID: \"266f85d9-996e-41a9-a83d-5451e74689b6\") " Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.198383 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/266f85d9-996e-41a9-a83d-5451e74689b6-fernet-keys\") pod \"266f85d9-996e-41a9-a83d-5451e74689b6\" (UID: \"266f85d9-996e-41a9-a83d-5451e74689b6\") " Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.198407 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3-combined-ca-bundle\") pod \"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3\" (UID: \"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3\") " Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.198430 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84s2j\" (UniqueName: \"kubernetes.io/projected/1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3-kube-api-access-84s2j\") pod \"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3\" (UID: \"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3\") " Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.198458 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/266f85d9-996e-41a9-a83d-5451e74689b6-scripts\") pod \"266f85d9-996e-41a9-a83d-5451e74689b6\" (UID: \"266f85d9-996e-41a9-a83d-5451e74689b6\") " Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.198536 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3-config-data\") pod \"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3\" (UID: \"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3\") " Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.198587 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3-httpd-run\") pod \"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3\" (UID: \"1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3\") " Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.202448 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3" (UID: "1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.208409 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3-logs" (OuterVolumeSpecName: "logs") pod "1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3" (UID: "1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.214046 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/266f85d9-996e-41a9-a83d-5451e74689b6-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "266f85d9-996e-41a9-a83d-5451e74689b6" (UID: "266f85d9-996e-41a9-a83d-5451e74689b6"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.216827 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/266f85d9-996e-41a9-a83d-5451e74689b6-scripts" (OuterVolumeSpecName: "scripts") pod "266f85d9-996e-41a9-a83d-5451e74689b6" (UID: "266f85d9-996e-41a9-a83d-5451e74689b6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.222549 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3-scripts" (OuterVolumeSpecName: "scripts") pod "1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3" (UID: "1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.222568 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/266f85d9-996e-41a9-a83d-5451e74689b6-kube-api-access-jfkhh" (OuterVolumeSpecName: "kube-api-access-jfkhh") pod "266f85d9-996e-41a9-a83d-5451e74689b6" (UID: "266f85d9-996e-41a9-a83d-5451e74689b6"). InnerVolumeSpecName "kube-api-access-jfkhh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.235489 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/266f85d9-996e-41a9-a83d-5451e74689b6-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "266f85d9-996e-41a9-a83d-5451e74689b6" (UID: "266f85d9-996e-41a9-a83d-5451e74689b6"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.235624 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3" (UID: "1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.240498 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3-kube-api-access-84s2j" (OuterVolumeSpecName: "kube-api-access-84s2j") pod "1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3" (UID: "1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3"). InnerVolumeSpecName "kube-api-access-84s2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.254517 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/266f85d9-996e-41a9-a83d-5451e74689b6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "266f85d9-996e-41a9-a83d-5451e74689b6" (UID: "266f85d9-996e-41a9-a83d-5451e74689b6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.272241 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3" (UID: "1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.294906 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/266f85d9-996e-41a9-a83d-5451e74689b6-config-data" (OuterVolumeSpecName: "config-data") pod "266f85d9-996e-41a9-a83d-5451e74689b6" (UID: "266f85d9-996e-41a9-a83d-5451e74689b6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.297681 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3-config-data" (OuterVolumeSpecName: "config-data") pod "1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3" (UID: "1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.302682 4787 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/266f85d9-996e-41a9-a83d-5451e74689b6-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.302726 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/266f85d9-996e-41a9-a83d-5451e74689b6-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.302736 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.302747 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/266f85d9-996e-41a9-a83d-5451e74689b6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.302756 4787 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.302793 4787 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.302803 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jfkhh\" (UniqueName: \"kubernetes.io/projected/266f85d9-996e-41a9-a83d-5451e74689b6-kube-api-access-jfkhh\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.302817 4787 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/266f85d9-996e-41a9-a83d-5451e74689b6-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.302827 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.302836 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84s2j\" (UniqueName: \"kubernetes.io/projected/1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3-kube-api-access-84s2j\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.302844 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/266f85d9-996e-41a9-a83d-5451e74689b6-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.302856 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.302867 4787 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.329401 4787 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.405850 4787 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.521882 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.547585 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.548014 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 09:52:22 crc kubenswrapper[4787]: E1001 09:52:22.550616 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3" containerName="glance-log" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.550642 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3" containerName="glance-log" Oct 01 09:52:22 crc kubenswrapper[4787]: E1001 09:52:22.550673 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3" containerName="glance-httpd" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.550679 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3" containerName="glance-httpd" Oct 01 09:52:22 crc kubenswrapper[4787]: E1001 09:52:22.550693 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="266f85d9-996e-41a9-a83d-5451e74689b6" containerName="keystone-bootstrap" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.550701 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="266f85d9-996e-41a9-a83d-5451e74689b6" containerName="keystone-bootstrap" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.550989 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3" containerName="glance-log" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.551007 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="266f85d9-996e-41a9-a83d-5451e74689b6" containerName="keystone-bootstrap" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.551031 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3" containerName="glance-httpd" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.554509 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.554627 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.558979 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.559062 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.621175 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1a020010-9967-4da2-9f0e-6ec5fc461137-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1a020010-9967-4da2-9f0e-6ec5fc461137\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.621321 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a020010-9967-4da2-9f0e-6ec5fc461137-logs\") pod \"glance-default-external-api-0\" (UID: \"1a020010-9967-4da2-9f0e-6ec5fc461137\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.621426 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a020010-9967-4da2-9f0e-6ec5fc461137-config-data\") pod \"glance-default-external-api-0\" (UID: \"1a020010-9967-4da2-9f0e-6ec5fc461137\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.621509 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkvtc\" (UniqueName: \"kubernetes.io/projected/1a020010-9967-4da2-9f0e-6ec5fc461137-kube-api-access-zkvtc\") pod \"glance-default-external-api-0\" (UID: \"1a020010-9967-4da2-9f0e-6ec5fc461137\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.621567 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a020010-9967-4da2-9f0e-6ec5fc461137-scripts\") pod \"glance-default-external-api-0\" (UID: \"1a020010-9967-4da2-9f0e-6ec5fc461137\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.621657 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a020010-9967-4da2-9f0e-6ec5fc461137-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"1a020010-9967-4da2-9f0e-6ec5fc461137\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.621997 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a020010-9967-4da2-9f0e-6ec5fc461137-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1a020010-9967-4da2-9f0e-6ec5fc461137\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.622087 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"1a020010-9967-4da2-9f0e-6ec5fc461137\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:22 crc kubenswrapper[4787]: E1001 09:52:22.706185 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:32a25ac44706b73bff04a89514177b1efd675f0442b295e225f0020555ca6350" Oct 01 09:52:22 crc kubenswrapper[4787]: E1001 09:52:22.706393 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:32a25ac44706b73bff04a89514177b1efd675f0442b295e225f0020555ca6350,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n665h67ch5d8h554h7chfbh56fh658h5d5h5f5h68bh64h675h674hf6h588h659h5fdh95h79h76h567h655h546h595hd8h5f6h79h55chf8h565h5c7q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cs95k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(9587d64e-33e3-45d2-8ee7-a776fcc60d88): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.724448 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1a020010-9967-4da2-9f0e-6ec5fc461137-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1a020010-9967-4da2-9f0e-6ec5fc461137\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.724514 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a020010-9967-4da2-9f0e-6ec5fc461137-logs\") pod \"glance-default-external-api-0\" (UID: \"1a020010-9967-4da2-9f0e-6ec5fc461137\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.724544 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a020010-9967-4da2-9f0e-6ec5fc461137-config-data\") pod \"glance-default-external-api-0\" (UID: \"1a020010-9967-4da2-9f0e-6ec5fc461137\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.724575 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkvtc\" (UniqueName: \"kubernetes.io/projected/1a020010-9967-4da2-9f0e-6ec5fc461137-kube-api-access-zkvtc\") pod \"glance-default-external-api-0\" (UID: \"1a020010-9967-4da2-9f0e-6ec5fc461137\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.724716 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a020010-9967-4da2-9f0e-6ec5fc461137-scripts\") pod \"glance-default-external-api-0\" (UID: \"1a020010-9967-4da2-9f0e-6ec5fc461137\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.725485 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a020010-9967-4da2-9f0e-6ec5fc461137-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"1a020010-9967-4da2-9f0e-6ec5fc461137\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.725516 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a020010-9967-4da2-9f0e-6ec5fc461137-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1a020010-9967-4da2-9f0e-6ec5fc461137\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.725547 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"1a020010-9967-4da2-9f0e-6ec5fc461137\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.725144 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a020010-9967-4da2-9f0e-6ec5fc461137-logs\") pod \"glance-default-external-api-0\" (UID: \"1a020010-9967-4da2-9f0e-6ec5fc461137\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.725857 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"1a020010-9967-4da2-9f0e-6ec5fc461137\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.725096 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1a020010-9967-4da2-9f0e-6ec5fc461137-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1a020010-9967-4da2-9f0e-6ec5fc461137\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.731640 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a020010-9967-4da2-9f0e-6ec5fc461137-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"1a020010-9967-4da2-9f0e-6ec5fc461137\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.731684 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a020010-9967-4da2-9f0e-6ec5fc461137-scripts\") pod \"glance-default-external-api-0\" (UID: \"1a020010-9967-4da2-9f0e-6ec5fc461137\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.732587 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a020010-9967-4da2-9f0e-6ec5fc461137-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1a020010-9967-4da2-9f0e-6ec5fc461137\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.732808 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a020010-9967-4da2-9f0e-6ec5fc461137-config-data\") pod \"glance-default-external-api-0\" (UID: \"1a020010-9967-4da2-9f0e-6ec5fc461137\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.744113 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkvtc\" (UniqueName: \"kubernetes.io/projected/1a020010-9967-4da2-9f0e-6ec5fc461137-kube-api-access-zkvtc\") pod \"glance-default-external-api-0\" (UID: \"1a020010-9967-4da2-9f0e-6ec5fc461137\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.754677 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"1a020010-9967-4da2-9f0e-6ec5fc461137\") " pod="openstack/glance-default-external-api-0" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.819341 4787 scope.go:117] "RemoveContainer" containerID="c980e4cea0aa90172278d8133be1046b80fde5491d7e385d078843d35aa6b323" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.849293 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7565dcfdb5-mcrm8" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.877060 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-58779b6b6f-rqbn7" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.878581 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-b48666669-4sqql" Oct 01 09:52:22 crc kubenswrapper[4787]: I1001 09:52:22.883053 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.012252 4787 scope.go:117] "RemoveContainer" containerID="16d8a5275751e42e56226fe37a99ccde7f83e12a427fc146cbb0a8833077315c" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.035033 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdpd2\" (UniqueName: \"kubernetes.io/projected/54c87b91-944a-4f8c-a556-21649f3b9b1b-kube-api-access-mdpd2\") pod \"54c87b91-944a-4f8c-a556-21649f3b9b1b\" (UID: \"54c87b91-944a-4f8c-a556-21649f3b9b1b\") " Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.035216 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3869928b-1191-433d-87bf-7187615b57e0-horizon-secret-key\") pod \"3869928b-1191-433d-87bf-7187615b57e0\" (UID: \"3869928b-1191-433d-87bf-7187615b57e0\") " Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.035265 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/54c87b91-944a-4f8c-a556-21649f3b9b1b-scripts\") pod \"54c87b91-944a-4f8c-a556-21649f3b9b1b\" (UID: \"54c87b91-944a-4f8c-a556-21649f3b9b1b\") " Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.035359 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/54c87b91-944a-4f8c-a556-21649f3b9b1b-horizon-secret-key\") pod \"54c87b91-944a-4f8c-a556-21649f3b9b1b\" (UID: \"54c87b91-944a-4f8c-a556-21649f3b9b1b\") " Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.035409 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3-config-data\") pod \"2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3\" (UID: \"2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3\") " Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.035438 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/54c87b91-944a-4f8c-a556-21649f3b9b1b-logs\") pod \"54c87b91-944a-4f8c-a556-21649f3b9b1b\" (UID: \"54c87b91-944a-4f8c-a556-21649f3b9b1b\") " Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.035469 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3869928b-1191-433d-87bf-7187615b57e0-config-data\") pod \"3869928b-1191-433d-87bf-7187615b57e0\" (UID: \"3869928b-1191-433d-87bf-7187615b57e0\") " Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.035511 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vncx9\" (UniqueName: \"kubernetes.io/projected/2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3-kube-api-access-vncx9\") pod \"2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3\" (UID: \"2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3\") " Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.035541 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3-scripts\") pod \"2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3\" (UID: \"2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3\") " Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.035565 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3-horizon-secret-key\") pod \"2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3\" (UID: \"2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3\") " Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.035596 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3869928b-1191-433d-87bf-7187615b57e0-logs\") pod \"3869928b-1191-433d-87bf-7187615b57e0\" (UID: \"3869928b-1191-433d-87bf-7187615b57e0\") " Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.035659 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3869928b-1191-433d-87bf-7187615b57e0-scripts\") pod \"3869928b-1191-433d-87bf-7187615b57e0\" (UID: \"3869928b-1191-433d-87bf-7187615b57e0\") " Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.035785 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/54c87b91-944a-4f8c-a556-21649f3b9b1b-config-data\") pod \"54c87b91-944a-4f8c-a556-21649f3b9b1b\" (UID: \"54c87b91-944a-4f8c-a556-21649f3b9b1b\") " Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.035836 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3-logs\") pod \"2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3\" (UID: \"2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3\") " Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.035912 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gvxbh\" (UniqueName: \"kubernetes.io/projected/3869928b-1191-433d-87bf-7187615b57e0-kube-api-access-gvxbh\") pod \"3869928b-1191-433d-87bf-7187615b57e0\" (UID: \"3869928b-1191-433d-87bf-7187615b57e0\") " Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.040756 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3869928b-1191-433d-87bf-7187615b57e0-logs" (OuterVolumeSpecName: "logs") pod "3869928b-1191-433d-87bf-7187615b57e0" (UID: "3869928b-1191-433d-87bf-7187615b57e0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.041473 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3-scripts" (OuterVolumeSpecName: "scripts") pod "2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3" (UID: "2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.041504 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54c87b91-944a-4f8c-a556-21649f3b9b1b-config-data" (OuterVolumeSpecName: "config-data") pod "54c87b91-944a-4f8c-a556-21649f3b9b1b" (UID: "54c87b91-944a-4f8c-a556-21649f3b9b1b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.042307 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3869928b-1191-433d-87bf-7187615b57e0-scripts" (OuterVolumeSpecName: "scripts") pod "3869928b-1191-433d-87bf-7187615b57e0" (UID: "3869928b-1191-433d-87bf-7187615b57e0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.043462 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54c87b91-944a-4f8c-a556-21649f3b9b1b-scripts" (OuterVolumeSpecName: "scripts") pod "54c87b91-944a-4f8c-a556-21649f3b9b1b" (UID: "54c87b91-944a-4f8c-a556-21649f3b9b1b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.043611 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54c87b91-944a-4f8c-a556-21649f3b9b1b-kube-api-access-mdpd2" (OuterVolumeSpecName: "kube-api-access-mdpd2") pod "54c87b91-944a-4f8c-a556-21649f3b9b1b" (UID: "54c87b91-944a-4f8c-a556-21649f3b9b1b"). InnerVolumeSpecName "kube-api-access-mdpd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.044008 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3-logs" (OuterVolumeSpecName: "logs") pod "2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3" (UID: "2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.044028 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54c87b91-944a-4f8c-a556-21649f3b9b1b-logs" (OuterVolumeSpecName: "logs") pod "54c87b91-944a-4f8c-a556-21649f3b9b1b" (UID: "54c87b91-944a-4f8c-a556-21649f3b9b1b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.045097 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3-config-data" (OuterVolumeSpecName: "config-data") pod "2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3" (UID: "2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.045333 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3869928b-1191-433d-87bf-7187615b57e0-kube-api-access-gvxbh" (OuterVolumeSpecName: "kube-api-access-gvxbh") pod "3869928b-1191-433d-87bf-7187615b57e0" (UID: "3869928b-1191-433d-87bf-7187615b57e0"). InnerVolumeSpecName "kube-api-access-gvxbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.045956 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3869928b-1191-433d-87bf-7187615b57e0-config-data" (OuterVolumeSpecName: "config-data") pod "3869928b-1191-433d-87bf-7187615b57e0" (UID: "3869928b-1191-433d-87bf-7187615b57e0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.047402 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3" (UID: "2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.048003 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3869928b-1191-433d-87bf-7187615b57e0-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "3869928b-1191-433d-87bf-7187615b57e0" (UID: "3869928b-1191-433d-87bf-7187615b57e0"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.048427 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3-kube-api-access-vncx9" (OuterVolumeSpecName: "kube-api-access-vncx9") pod "2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3" (UID: "2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3"). InnerVolumeSpecName "kube-api-access-vncx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.052889 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54c87b91-944a-4f8c-a556-21649f3b9b1b-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "54c87b91-944a-4f8c-a556-21649f3b9b1b" (UID: "54c87b91-944a-4f8c-a556-21649f3b9b1b"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.139396 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/54c87b91-944a-4f8c-a556-21649f3b9b1b-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.139794 4787 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.139807 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gvxbh\" (UniqueName: \"kubernetes.io/projected/3869928b-1191-433d-87bf-7187615b57e0-kube-api-access-gvxbh\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.139819 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdpd2\" (UniqueName: \"kubernetes.io/projected/54c87b91-944a-4f8c-a556-21649f3b9b1b-kube-api-access-mdpd2\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.139830 4787 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3869928b-1191-433d-87bf-7187615b57e0-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.139843 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/54c87b91-944a-4f8c-a556-21649f3b9b1b-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.139851 4787 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/54c87b91-944a-4f8c-a556-21649f3b9b1b-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.139861 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.139870 4787 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/54c87b91-944a-4f8c-a556-21649f3b9b1b-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.139879 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3869928b-1191-433d-87bf-7187615b57e0-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.139889 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vncx9\" (UniqueName: \"kubernetes.io/projected/2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3-kube-api-access-vncx9\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.139897 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.139905 4787 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.139913 4787 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3869928b-1191-433d-87bf-7187615b57e0-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.139922 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3869928b-1191-433d-87bf-7187615b57e0-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.162581 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7565dcfdb5-mcrm8" event={"ID":"54c87b91-944a-4f8c-a556-21649f3b9b1b","Type":"ContainerDied","Data":"8fe9847b510592d21529b6e9e707acfc71ff4f1873c32df5efe1b30aa84f7c07"} Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.162773 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7565dcfdb5-mcrm8" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.192393 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-b48666669-4sqql" event={"ID":"3869928b-1191-433d-87bf-7187615b57e0","Type":"ContainerDied","Data":"826baf402d328a062c18d66aeae1e510c3b2433b269b21aa1da7daceb12f0b51"} Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.192579 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-b48666669-4sqql" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.216213 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-58779b6b6f-rqbn7" event={"ID":"2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3","Type":"ContainerDied","Data":"cea04ca942277b341a7475576791e4622eb8bc4d8d205d7cbc5befacc79d6de3"} Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.216327 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-58779b6b6f-rqbn7" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.219564 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-tk8w8"] Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.221698 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c6d5d5bd7-4mbvj" event={"ID":"bbade729-c3cb-447b-bdd7-251fbc80c454","Type":"ContainerDied","Data":"686aeae305b7a63e69cd2cdf66eb516ba5c4667697ac7753b580e525d09e7cea"} Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.221737 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="686aeae305b7a63e69cd2cdf66eb516ba5c4667697ac7753b580e525d09e7cea" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.225200 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-tk8w8"] Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.294383 4787 scope.go:117] "RemoveContainer" containerID="bbd0a2d96558ec05c6a18fc511729c45b4afffab39efef62e9345bc1144c0d8f" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.337625 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c6d5d5bd7-4mbvj" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.341372 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdp72\" (UniqueName: \"kubernetes.io/projected/bbade729-c3cb-447b-bdd7-251fbc80c454-kube-api-access-wdp72\") pod \"bbade729-c3cb-447b-bdd7-251fbc80c454\" (UID: \"bbade729-c3cb-447b-bdd7-251fbc80c454\") " Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.341539 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bbade729-c3cb-447b-bdd7-251fbc80c454-config\") pod \"bbade729-c3cb-447b-bdd7-251fbc80c454\" (UID: \"bbade729-c3cb-447b-bdd7-251fbc80c454\") " Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.343497 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bbade729-c3cb-447b-bdd7-251fbc80c454-ovsdbserver-sb\") pod \"bbade729-c3cb-447b-bdd7-251fbc80c454\" (UID: \"bbade729-c3cb-447b-bdd7-251fbc80c454\") " Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.343661 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bbade729-c3cb-447b-bdd7-251fbc80c454-ovsdbserver-nb\") pod \"bbade729-c3cb-447b-bdd7-251fbc80c454\" (UID: \"bbade729-c3cb-447b-bdd7-251fbc80c454\") " Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.343733 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bbade729-c3cb-447b-bdd7-251fbc80c454-dns-svc\") pod \"bbade729-c3cb-447b-bdd7-251fbc80c454\" (UID: \"bbade729-c3cb-447b-bdd7-251fbc80c454\") " Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.379562 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbade729-c3cb-447b-bdd7-251fbc80c454-kube-api-access-wdp72" (OuterVolumeSpecName: "kube-api-access-wdp72") pod "bbade729-c3cb-447b-bdd7-251fbc80c454" (UID: "bbade729-c3cb-447b-bdd7-251fbc80c454"). InnerVolumeSpecName "kube-api-access-wdp72". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.411757 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7565dcfdb5-mcrm8"] Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.421255 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7565dcfdb5-mcrm8"] Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.427884 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-dx2gc"] Oct 01 09:52:23 crc kubenswrapper[4787]: E1001 09:52:23.428370 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbade729-c3cb-447b-bdd7-251fbc80c454" containerName="init" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.428392 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbade729-c3cb-447b-bdd7-251fbc80c454" containerName="init" Oct 01 09:52:23 crc kubenswrapper[4787]: E1001 09:52:23.428447 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbade729-c3cb-447b-bdd7-251fbc80c454" containerName="dnsmasq-dns" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.428456 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbade729-c3cb-447b-bdd7-251fbc80c454" containerName="dnsmasq-dns" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.428630 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbade729-c3cb-447b-bdd7-251fbc80c454" containerName="dnsmasq-dns" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.429306 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dx2gc" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.431898 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.433126 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.433370 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.433708 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-qlb5k" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.435905 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-dx2gc"] Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.448204 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/345ecb87-2ef8-4829-8137-156dec6d0e80-config-data\") pod \"keystone-bootstrap-dx2gc\" (UID: \"345ecb87-2ef8-4829-8137-156dec6d0e80\") " pod="openstack/keystone-bootstrap-dx2gc" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.448540 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/345ecb87-2ef8-4829-8137-156dec6d0e80-fernet-keys\") pod \"keystone-bootstrap-dx2gc\" (UID: \"345ecb87-2ef8-4829-8137-156dec6d0e80\") " pod="openstack/keystone-bootstrap-dx2gc" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.448735 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvfms\" (UniqueName: \"kubernetes.io/projected/345ecb87-2ef8-4829-8137-156dec6d0e80-kube-api-access-jvfms\") pod \"keystone-bootstrap-dx2gc\" (UID: \"345ecb87-2ef8-4829-8137-156dec6d0e80\") " pod="openstack/keystone-bootstrap-dx2gc" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.448934 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/345ecb87-2ef8-4829-8137-156dec6d0e80-combined-ca-bundle\") pod \"keystone-bootstrap-dx2gc\" (UID: \"345ecb87-2ef8-4829-8137-156dec6d0e80\") " pod="openstack/keystone-bootstrap-dx2gc" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.449455 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/345ecb87-2ef8-4829-8137-156dec6d0e80-credential-keys\") pod \"keystone-bootstrap-dx2gc\" (UID: \"345ecb87-2ef8-4829-8137-156dec6d0e80\") " pod="openstack/keystone-bootstrap-dx2gc" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.449659 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/345ecb87-2ef8-4829-8137-156dec6d0e80-scripts\") pod \"keystone-bootstrap-dx2gc\" (UID: \"345ecb87-2ef8-4829-8137-156dec6d0e80\") " pod="openstack/keystone-bootstrap-dx2gc" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.449990 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdp72\" (UniqueName: \"kubernetes.io/projected/bbade729-c3cb-447b-bdd7-251fbc80c454-kube-api-access-wdp72\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.490134 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-b48666669-4sqql"] Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.511097 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-b48666669-4sqql"] Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.524677 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-58779b6b6f-rqbn7"] Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.540727 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bbade729-c3cb-447b-bdd7-251fbc80c454-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bbade729-c3cb-447b-bdd7-251fbc80c454" (UID: "bbade729-c3cb-447b-bdd7-251fbc80c454"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.542519 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bbade729-c3cb-447b-bdd7-251fbc80c454-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bbade729-c3cb-447b-bdd7-251fbc80c454" (UID: "bbade729-c3cb-447b-bdd7-251fbc80c454"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.545804 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bbade729-c3cb-447b-bdd7-251fbc80c454-config" (OuterVolumeSpecName: "config") pod "bbade729-c3cb-447b-bdd7-251fbc80c454" (UID: "bbade729-c3cb-447b-bdd7-251fbc80c454"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.552942 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/345ecb87-2ef8-4829-8137-156dec6d0e80-scripts\") pod \"keystone-bootstrap-dx2gc\" (UID: \"345ecb87-2ef8-4829-8137-156dec6d0e80\") " pod="openstack/keystone-bootstrap-dx2gc" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.553021 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/345ecb87-2ef8-4829-8137-156dec6d0e80-config-data\") pod \"keystone-bootstrap-dx2gc\" (UID: \"345ecb87-2ef8-4829-8137-156dec6d0e80\") " pod="openstack/keystone-bootstrap-dx2gc" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.553098 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/345ecb87-2ef8-4829-8137-156dec6d0e80-fernet-keys\") pod \"keystone-bootstrap-dx2gc\" (UID: \"345ecb87-2ef8-4829-8137-156dec6d0e80\") " pod="openstack/keystone-bootstrap-dx2gc" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.553162 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvfms\" (UniqueName: \"kubernetes.io/projected/345ecb87-2ef8-4829-8137-156dec6d0e80-kube-api-access-jvfms\") pod \"keystone-bootstrap-dx2gc\" (UID: \"345ecb87-2ef8-4829-8137-156dec6d0e80\") " pod="openstack/keystone-bootstrap-dx2gc" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.553199 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/345ecb87-2ef8-4829-8137-156dec6d0e80-combined-ca-bundle\") pod \"keystone-bootstrap-dx2gc\" (UID: \"345ecb87-2ef8-4829-8137-156dec6d0e80\") " pod="openstack/keystone-bootstrap-dx2gc" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.553327 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/345ecb87-2ef8-4829-8137-156dec6d0e80-credential-keys\") pod \"keystone-bootstrap-dx2gc\" (UID: \"345ecb87-2ef8-4829-8137-156dec6d0e80\") " pod="openstack/keystone-bootstrap-dx2gc" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.553439 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bbade729-c3cb-447b-bdd7-251fbc80c454-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.553462 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bbade729-c3cb-447b-bdd7-251fbc80c454-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.553496 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bbade729-c3cb-447b-bdd7-251fbc80c454-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.558104 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-58779b6b6f-rqbn7"] Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.559351 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/345ecb87-2ef8-4829-8137-156dec6d0e80-fernet-keys\") pod \"keystone-bootstrap-dx2gc\" (UID: \"345ecb87-2ef8-4829-8137-156dec6d0e80\") " pod="openstack/keystone-bootstrap-dx2gc" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.559365 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/345ecb87-2ef8-4829-8137-156dec6d0e80-combined-ca-bundle\") pod \"keystone-bootstrap-dx2gc\" (UID: \"345ecb87-2ef8-4829-8137-156dec6d0e80\") " pod="openstack/keystone-bootstrap-dx2gc" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.561635 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/345ecb87-2ef8-4829-8137-156dec6d0e80-scripts\") pod \"keystone-bootstrap-dx2gc\" (UID: \"345ecb87-2ef8-4829-8137-156dec6d0e80\") " pod="openstack/keystone-bootstrap-dx2gc" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.565694 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/345ecb87-2ef8-4829-8137-156dec6d0e80-config-data\") pod \"keystone-bootstrap-dx2gc\" (UID: \"345ecb87-2ef8-4829-8137-156dec6d0e80\") " pod="openstack/keystone-bootstrap-dx2gc" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.570735 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/345ecb87-2ef8-4829-8137-156dec6d0e80-credential-keys\") pod \"keystone-bootstrap-dx2gc\" (UID: \"345ecb87-2ef8-4829-8137-156dec6d0e80\") " pod="openstack/keystone-bootstrap-dx2gc" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.578580 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvfms\" (UniqueName: \"kubernetes.io/projected/345ecb87-2ef8-4829-8137-156dec6d0e80-kube-api-access-jvfms\") pod \"keystone-bootstrap-dx2gc\" (UID: \"345ecb87-2ef8-4829-8137-156dec6d0e80\") " pod="openstack/keystone-bootstrap-dx2gc" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.591446 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bbade729-c3cb-447b-bdd7-251fbc80c454-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bbade729-c3cb-447b-bdd7-251fbc80c454" (UID: "bbade729-c3cb-447b-bdd7-251fbc80c454"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.655836 4787 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bbade729-c3cb-447b-bdd7-251fbc80c454-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.732038 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-819c-account-create-86wn2"] Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.738863 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-cf80-account-create-zz7cn"] Oct 01 09:52:23 crc kubenswrapper[4787]: W1001 09:52:23.742773 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod22964b11_e82b_4dfa_ac42_70005cbb5caf.slice/crio-782cf551a50bf36d2d7526e81001ef11f6371dbee3cacfff90bc338c02b8db26 WatchSource:0}: Error finding container 782cf551a50bf36d2d7526e81001ef11f6371dbee3cacfff90bc338c02b8db26: Status 404 returned error can't find the container with id 782cf551a50bf36d2d7526e81001ef11f6371dbee3cacfff90bc338c02b8db26 Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.761435 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dx2gc" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.763735 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.849420 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-eafd-account-create-h97jm"] Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.863165 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-67879968b-d5598"] Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.872296 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7d65d5957b-44rtv"] Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.881648 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:52:23 crc kubenswrapper[4787]: I1001 09:52:23.920293 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 01 09:52:24 crc kubenswrapper[4787]: I1001 09:52:24.081381 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 09:52:24 crc kubenswrapper[4787]: W1001 09:52:24.114460 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a020010_9967_4da2_9f0e_6ec5fc461137.slice/crio-5079c9b6b1b9bcfaac45797acdb2dcebf8fc508abe65ba630e5336fd6866ddc9 WatchSource:0}: Error finding container 5079c9b6b1b9bcfaac45797acdb2dcebf8fc508abe65ba630e5336fd6866ddc9: Status 404 returned error can't find the container with id 5079c9b6b1b9bcfaac45797acdb2dcebf8fc508abe65ba630e5336fd6866ddc9 Oct 01 09:52:24 crc kubenswrapper[4787]: I1001 09:52:24.234990 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"556ee9b9-8f10-4de4-9a97-1d3327f885f1","Type":"ContainerStarted","Data":"9735648a53d595db56c29fc3e5333ec0d4ec4e25353b8dbce92180cbbefc08bb"} Oct 01 09:52:24 crc kubenswrapper[4787]: I1001 09:52:24.237408 4787 generic.go:334] "Generic (PLEG): container finished" podID="22964b11-e82b-4dfa-ac42-70005cbb5caf" containerID="81db296f8f18bdde1e19f761178b08968998618eefe18e165ce27d4659f5cfd9" exitCode=0 Oct 01 09:52:24 crc kubenswrapper[4787]: I1001 09:52:24.237467 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-819c-account-create-86wn2" event={"ID":"22964b11-e82b-4dfa-ac42-70005cbb5caf","Type":"ContainerDied","Data":"81db296f8f18bdde1e19f761178b08968998618eefe18e165ce27d4659f5cfd9"} Oct 01 09:52:24 crc kubenswrapper[4787]: I1001 09:52:24.237484 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-819c-account-create-86wn2" event={"ID":"22964b11-e82b-4dfa-ac42-70005cbb5caf","Type":"ContainerStarted","Data":"782cf551a50bf36d2d7526e81001ef11f6371dbee3cacfff90bc338c02b8db26"} Oct 01 09:52:24 crc kubenswrapper[4787]: I1001 09:52:24.238788 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1a020010-9967-4da2-9f0e-6ec5fc461137","Type":"ContainerStarted","Data":"5079c9b6b1b9bcfaac45797acdb2dcebf8fc508abe65ba630e5336fd6866ddc9"} Oct 01 09:52:24 crc kubenswrapper[4787]: I1001 09:52:24.245219 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3432bc16-99b4-431a-aeb1-600e826bbc3e","Type":"ContainerStarted","Data":"de5e35e2fec13b600c1797f0db4642bfa22d5dd8b00a10f48fd2744a9fb963d3"} Oct 01 09:52:24 crc kubenswrapper[4787]: I1001 09:52:24.245311 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3432bc16-99b4-431a-aeb1-600e826bbc3e","Type":"ContainerStarted","Data":"2120a9ff1b20f9ed764530d903a918b711842676e72df28f48ed304eab3820ab"} Oct 01 09:52:24 crc kubenswrapper[4787]: I1001 09:52:24.245328 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3432bc16-99b4-431a-aeb1-600e826bbc3e","Type":"ContainerStarted","Data":"075f4fc837ec4a0a053186f8512986b479b30b1d7acc5e29e4d15a1fbd0520fc"} Oct 01 09:52:24 crc kubenswrapper[4787]: I1001 09:52:24.247178 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-eafd-account-create-h97jm" event={"ID":"bcadc39c-6259-49a9-9f5f-38a80545b5a4","Type":"ContainerStarted","Data":"27779446f41934986444ab73ee658c5bc366fdaef5b2a0c0465ae75308819735"} Oct 01 09:52:24 crc kubenswrapper[4787]: I1001 09:52:24.252114 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d65d5957b-44rtv" event={"ID":"db1d14ba-8d00-4508-9c89-fdadb562ad0d","Type":"ContainerStarted","Data":"39eae6a601ed10f7c4fe7ca4d85fa9ead5fbe5f63a02a69c0d4b5f2eaf691c3d"} Oct 01 09:52:24 crc kubenswrapper[4787]: I1001 09:52:24.257112 4787 generic.go:334] "Generic (PLEG): container finished" podID="54517204-c321-4959-851b-9c642f07a7a7" containerID="72eab7966f013f105d29b360d7b6c98c01ddfc7ab40598ddead1f1c8e77924f2" exitCode=0 Oct 01 09:52:24 crc kubenswrapper[4787]: I1001 09:52:24.257178 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-cf80-account-create-zz7cn" event={"ID":"54517204-c321-4959-851b-9c642f07a7a7","Type":"ContainerDied","Data":"72eab7966f013f105d29b360d7b6c98c01ddfc7ab40598ddead1f1c8e77924f2"} Oct 01 09:52:24 crc kubenswrapper[4787]: I1001 09:52:24.257255 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-cf80-account-create-zz7cn" event={"ID":"54517204-c321-4959-851b-9c642f07a7a7","Type":"ContainerStarted","Data":"9df69eee08bc28115201dadbabe32eb21a185972608a27d6bbc492fbc7408c21"} Oct 01 09:52:24 crc kubenswrapper[4787]: I1001 09:52:24.258653 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67879968b-d5598" event={"ID":"7983e213-c317-42b2-8944-85bdb78dac8f","Type":"ContainerStarted","Data":"b68c4125c2c2e1d0824ffc24ec8de859b0016496a3588a9f02f3ba487eb043ad"} Oct 01 09:52:24 crc kubenswrapper[4787]: I1001 09:52:24.260416 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c6d5d5bd7-4mbvj" Oct 01 09:52:24 crc kubenswrapper[4787]: I1001 09:52:24.260718 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-slvbz" event={"ID":"6af63266-547b-4537-9290-338b2c0a2d73","Type":"ContainerStarted","Data":"b80433841686fb3ebfee3df8efba794169f6019700338907134cc87d5328b4e4"} Oct 01 09:52:24 crc kubenswrapper[4787]: I1001 09:52:24.305501 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-slvbz" podStartSLOduration=3.381927448 podStartE2EDuration="23.305477929s" podCreationTimestamp="2025-10-01 09:52:01 +0000 UTC" firstStartedPulling="2025-10-01 09:52:02.82281835 +0000 UTC m=+954.937962507" lastFinishedPulling="2025-10-01 09:52:22.746368841 +0000 UTC m=+974.861512988" observedRunningTime="2025-10-01 09:52:24.296231284 +0000 UTC m=+976.411375431" watchObservedRunningTime="2025-10-01 09:52:24.305477929 +0000 UTC m=+976.420622086" Oct 01 09:52:24 crc kubenswrapper[4787]: I1001 09:52:24.317739 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-dx2gc"] Oct 01 09:52:24 crc kubenswrapper[4787]: I1001 09:52:24.352752 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c6d5d5bd7-4mbvj"] Oct 01 09:52:24 crc kubenswrapper[4787]: I1001 09:52:24.360876 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6c6d5d5bd7-4mbvj"] Oct 01 09:52:24 crc kubenswrapper[4787]: I1001 09:52:24.534945 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3" path="/var/lib/kubelet/pods/1efe8cbc-e1b0-45cf-9e12-f5b92766bbf3/volumes" Oct 01 09:52:24 crc kubenswrapper[4787]: I1001 09:52:24.536254 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="266f85d9-996e-41a9-a83d-5451e74689b6" path="/var/lib/kubelet/pods/266f85d9-996e-41a9-a83d-5451e74689b6/volumes" Oct 01 09:52:24 crc kubenswrapper[4787]: I1001 09:52:24.536875 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3" path="/var/lib/kubelet/pods/2f73fe98-1f18-4c6d-9868-bd3dcd7ac8d3/volumes" Oct 01 09:52:24 crc kubenswrapper[4787]: I1001 09:52:24.537927 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3869928b-1191-433d-87bf-7187615b57e0" path="/var/lib/kubelet/pods/3869928b-1191-433d-87bf-7187615b57e0/volumes" Oct 01 09:52:24 crc kubenswrapper[4787]: I1001 09:52:24.538429 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54c87b91-944a-4f8c-a556-21649f3b9b1b" path="/var/lib/kubelet/pods/54c87b91-944a-4f8c-a556-21649f3b9b1b/volumes" Oct 01 09:52:24 crc kubenswrapper[4787]: I1001 09:52:24.538993 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbade729-c3cb-447b-bdd7-251fbc80c454" path="/var/lib/kubelet/pods/bbade729-c3cb-447b-bdd7-251fbc80c454/volumes" Oct 01 09:52:25 crc kubenswrapper[4787]: I1001 09:52:25.272127 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dx2gc" event={"ID":"345ecb87-2ef8-4829-8137-156dec6d0e80","Type":"ContainerStarted","Data":"0657daa702ef6496c3116fb6a7456ba0952ba05d4954c24ac71b5045b3e0c13b"} Oct 01 09:52:25 crc kubenswrapper[4787]: I1001 09:52:25.272676 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dx2gc" event={"ID":"345ecb87-2ef8-4829-8137-156dec6d0e80","Type":"ContainerStarted","Data":"cee82fc6e04215873beb43900300afa3baed5299bcf23d62d1c87f1deae39f6c"} Oct 01 09:52:25 crc kubenswrapper[4787]: I1001 09:52:25.278715 4787 generic.go:334] "Generic (PLEG): container finished" podID="bcadc39c-6259-49a9-9f5f-38a80545b5a4" containerID="b46dfa12aa0b9ae2977caa8b3df6e30cf29ff756fb395e9e390f056bcced7b29" exitCode=0 Oct 01 09:52:25 crc kubenswrapper[4787]: I1001 09:52:25.278781 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-eafd-account-create-h97jm" event={"ID":"bcadc39c-6259-49a9-9f5f-38a80545b5a4","Type":"ContainerDied","Data":"b46dfa12aa0b9ae2977caa8b3df6e30cf29ff756fb395e9e390f056bcced7b29"} Oct 01 09:52:25 crc kubenswrapper[4787]: I1001 09:52:25.281411 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67879968b-d5598" event={"ID":"7983e213-c317-42b2-8944-85bdb78dac8f","Type":"ContainerStarted","Data":"37b1a9eaf9bc59c2abf28488f9ac98956d1e78ecc9bdf3f5d70a1af066d4e549"} Oct 01 09:52:25 crc kubenswrapper[4787]: I1001 09:52:25.285916 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"556ee9b9-8f10-4de4-9a97-1d3327f885f1","Type":"ContainerStarted","Data":"9f3da2459ceb85b1e397ec6bdc5977d7ffaadeed925c8538223d712b95ad33c8"} Oct 01 09:52:25 crc kubenswrapper[4787]: I1001 09:52:25.290916 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-dx2gc" podStartSLOduration=2.290900801 podStartE2EDuration="2.290900801s" podCreationTimestamp="2025-10-01 09:52:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:52:25.288164851 +0000 UTC m=+977.403309008" watchObservedRunningTime="2025-10-01 09:52:25.290900801 +0000 UTC m=+977.406044958" Oct 01 09:52:25 crc kubenswrapper[4787]: I1001 09:52:25.301373 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3432bc16-99b4-431a-aeb1-600e826bbc3e","Type":"ContainerStarted","Data":"fac27356bb9cbc88386a1430f694f742b087d5db6708b210f5ecc4b5c7d744da"} Oct 01 09:52:25 crc kubenswrapper[4787]: I1001 09:52:25.308379 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9587d64e-33e3-45d2-8ee7-a776fcc60d88","Type":"ContainerStarted","Data":"7564cf5c7eb4e3180d068e49673d5fb7e6cd1d11878d243edfaee1b52031f33d"} Oct 01 09:52:25 crc kubenswrapper[4787]: I1001 09:52:25.328626 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d65d5957b-44rtv" event={"ID":"db1d14ba-8d00-4508-9c89-fdadb562ad0d","Type":"ContainerStarted","Data":"8724dff7866faff6d5ed6e0857b0561cb2e7df692abe6dde5869a7efc657b5ca"} Oct 01 09:52:25 crc kubenswrapper[4787]: I1001 09:52:25.716328 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-cf80-account-create-zz7cn" Oct 01 09:52:25 crc kubenswrapper[4787]: I1001 09:52:25.801310 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-819c-account-create-86wn2" Oct 01 09:52:25 crc kubenswrapper[4787]: I1001 09:52:25.822115 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjbdh\" (UniqueName: \"kubernetes.io/projected/54517204-c321-4959-851b-9c642f07a7a7-kube-api-access-zjbdh\") pod \"54517204-c321-4959-851b-9c642f07a7a7\" (UID: \"54517204-c321-4959-851b-9c642f07a7a7\") " Oct 01 09:52:25 crc kubenswrapper[4787]: I1001 09:52:25.828000 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54517204-c321-4959-851b-9c642f07a7a7-kube-api-access-zjbdh" (OuterVolumeSpecName: "kube-api-access-zjbdh") pod "54517204-c321-4959-851b-9c642f07a7a7" (UID: "54517204-c321-4959-851b-9c642f07a7a7"). InnerVolumeSpecName "kube-api-access-zjbdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:52:25 crc kubenswrapper[4787]: I1001 09:52:25.924259 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dmmj\" (UniqueName: \"kubernetes.io/projected/22964b11-e82b-4dfa-ac42-70005cbb5caf-kube-api-access-2dmmj\") pod \"22964b11-e82b-4dfa-ac42-70005cbb5caf\" (UID: \"22964b11-e82b-4dfa-ac42-70005cbb5caf\") " Oct 01 09:52:25 crc kubenswrapper[4787]: I1001 09:52:25.925018 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjbdh\" (UniqueName: \"kubernetes.io/projected/54517204-c321-4959-851b-9c642f07a7a7-kube-api-access-zjbdh\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:25 crc kubenswrapper[4787]: I1001 09:52:25.929160 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22964b11-e82b-4dfa-ac42-70005cbb5caf-kube-api-access-2dmmj" (OuterVolumeSpecName: "kube-api-access-2dmmj") pod "22964b11-e82b-4dfa-ac42-70005cbb5caf" (UID: "22964b11-e82b-4dfa-ac42-70005cbb5caf"). InnerVolumeSpecName "kube-api-access-2dmmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:52:26 crc kubenswrapper[4787]: I1001 09:52:26.026217 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dmmj\" (UniqueName: \"kubernetes.io/projected/22964b11-e82b-4dfa-ac42-70005cbb5caf-kube-api-access-2dmmj\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:26 crc kubenswrapper[4787]: I1001 09:52:26.345649 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67879968b-d5598" event={"ID":"7983e213-c317-42b2-8944-85bdb78dac8f","Type":"ContainerStarted","Data":"c363923f03c8431731ce61fb7088f6a51538ab7d45b2f5fb59d3d2d245ae02f0"} Oct 01 09:52:26 crc kubenswrapper[4787]: I1001 09:52:26.347406 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"556ee9b9-8f10-4de4-9a97-1d3327f885f1","Type":"ContainerStarted","Data":"1c20a8d94201178c836396310098095ed405417f7dcd3ea216a7d70f52076897"} Oct 01 09:52:26 crc kubenswrapper[4787]: I1001 09:52:26.347574 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="556ee9b9-8f10-4de4-9a97-1d3327f885f1" containerName="glance-log" containerID="cri-o://9f3da2459ceb85b1e397ec6bdc5977d7ffaadeed925c8538223d712b95ad33c8" gracePeriod=30 Oct 01 09:52:26 crc kubenswrapper[4787]: I1001 09:52:26.347611 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="556ee9b9-8f10-4de4-9a97-1d3327f885f1" containerName="glance-httpd" containerID="cri-o://1c20a8d94201178c836396310098095ed405417f7dcd3ea216a7d70f52076897" gracePeriod=30 Oct 01 09:52:26 crc kubenswrapper[4787]: I1001 09:52:26.352044 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-819c-account-create-86wn2" event={"ID":"22964b11-e82b-4dfa-ac42-70005cbb5caf","Type":"ContainerDied","Data":"782cf551a50bf36d2d7526e81001ef11f6371dbee3cacfff90bc338c02b8db26"} Oct 01 09:52:26 crc kubenswrapper[4787]: I1001 09:52:26.352097 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="782cf551a50bf36d2d7526e81001ef11f6371dbee3cacfff90bc338c02b8db26" Oct 01 09:52:26 crc kubenswrapper[4787]: I1001 09:52:26.352052 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-819c-account-create-86wn2" Oct 01 09:52:26 crc kubenswrapper[4787]: I1001 09:52:26.356612 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1a020010-9967-4da2-9f0e-6ec5fc461137","Type":"ContainerStarted","Data":"ddc800b2f9d11dcc930a8b39989f2bbc9f97a490a9ba0389ce5ceab59fe9641b"} Oct 01 09:52:26 crc kubenswrapper[4787]: I1001 09:52:26.356668 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1a020010-9967-4da2-9f0e-6ec5fc461137","Type":"ContainerStarted","Data":"c1bdee63e5aab6c9222b16cd9ff592390efcad78109917777cf0033f2a45414c"} Oct 01 09:52:26 crc kubenswrapper[4787]: I1001 09:52:26.358329 4787 generic.go:334] "Generic (PLEG): container finished" podID="6af63266-547b-4537-9290-338b2c0a2d73" containerID="b80433841686fb3ebfee3df8efba794169f6019700338907134cc87d5328b4e4" exitCode=0 Oct 01 09:52:26 crc kubenswrapper[4787]: I1001 09:52:26.358414 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-slvbz" event={"ID":"6af63266-547b-4537-9290-338b2c0a2d73","Type":"ContainerDied","Data":"b80433841686fb3ebfee3df8efba794169f6019700338907134cc87d5328b4e4"} Oct 01 09:52:26 crc kubenswrapper[4787]: I1001 09:52:26.360686 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d65d5957b-44rtv" event={"ID":"db1d14ba-8d00-4508-9c89-fdadb562ad0d","Type":"ContainerStarted","Data":"e184b7d9d7c394aa26fd1bd423a35ec5624001d544d8ff4f828fc61897aa23d7"} Oct 01 09:52:26 crc kubenswrapper[4787]: I1001 09:52:26.367370 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-cf80-account-create-zz7cn" Oct 01 09:52:26 crc kubenswrapper[4787]: I1001 09:52:26.367365 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-cf80-account-create-zz7cn" event={"ID":"54517204-c321-4959-851b-9c642f07a7a7","Type":"ContainerDied","Data":"9df69eee08bc28115201dadbabe32eb21a185972608a27d6bbc492fbc7408c21"} Oct 01 09:52:26 crc kubenswrapper[4787]: I1001 09:52:26.367499 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9df69eee08bc28115201dadbabe32eb21a185972608a27d6bbc492fbc7408c21" Oct 01 09:52:26 crc kubenswrapper[4787]: I1001 09:52:26.379517 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-67879968b-d5598" podStartSLOduration=13.615188851 podStartE2EDuration="14.379496354s" podCreationTimestamp="2025-10-01 09:52:12 +0000 UTC" firstStartedPulling="2025-10-01 09:52:23.913311049 +0000 UTC m=+976.028455206" lastFinishedPulling="2025-10-01 09:52:24.677618552 +0000 UTC m=+976.792762709" observedRunningTime="2025-10-01 09:52:26.377862493 +0000 UTC m=+978.493006640" watchObservedRunningTime="2025-10-01 09:52:26.379496354 +0000 UTC m=+978.494640511" Oct 01 09:52:26 crc kubenswrapper[4787]: I1001 09:52:26.408982 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7d65d5957b-44rtv" podStartSLOduration=13.631182136 podStartE2EDuration="14.4089581s" podCreationTimestamp="2025-10-01 09:52:12 +0000 UTC" firstStartedPulling="2025-10-01 09:52:23.917125995 +0000 UTC m=+976.032270142" lastFinishedPulling="2025-10-01 09:52:24.694901949 +0000 UTC m=+976.810046106" observedRunningTime="2025-10-01 09:52:26.403738468 +0000 UTC m=+978.518882645" watchObservedRunningTime="2025-10-01 09:52:26.4089581 +0000 UTC m=+978.524102257" Oct 01 09:52:26 crc kubenswrapper[4787]: I1001 09:52:26.456311 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=17.456285899 podStartE2EDuration="17.456285899s" podCreationTimestamp="2025-10-01 09:52:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:52:26.453261812 +0000 UTC m=+978.568405979" watchObservedRunningTime="2025-10-01 09:52:26.456285899 +0000 UTC m=+978.571430066" Oct 01 09:52:26 crc kubenswrapper[4787]: I1001 09:52:26.485688 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.485662583 podStartE2EDuration="4.485662583s" podCreationTimestamp="2025-10-01 09:52:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:52:26.480233715 +0000 UTC m=+978.595377872" watchObservedRunningTime="2025-10-01 09:52:26.485662583 +0000 UTC m=+978.600806740" Oct 01 09:52:26 crc kubenswrapper[4787]: I1001 09:52:26.765462 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-eafd-account-create-h97jm" Oct 01 09:52:26 crc kubenswrapper[4787]: I1001 09:52:26.841037 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96dcm\" (UniqueName: \"kubernetes.io/projected/bcadc39c-6259-49a9-9f5f-38a80545b5a4-kube-api-access-96dcm\") pod \"bcadc39c-6259-49a9-9f5f-38a80545b5a4\" (UID: \"bcadc39c-6259-49a9-9f5f-38a80545b5a4\") " Oct 01 09:52:26 crc kubenswrapper[4787]: I1001 09:52:26.870808 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcadc39c-6259-49a9-9f5f-38a80545b5a4-kube-api-access-96dcm" (OuterVolumeSpecName: "kube-api-access-96dcm") pod "bcadc39c-6259-49a9-9f5f-38a80545b5a4" (UID: "bcadc39c-6259-49a9-9f5f-38a80545b5a4"). InnerVolumeSpecName "kube-api-access-96dcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:52:26 crc kubenswrapper[4787]: I1001 09:52:26.943799 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96dcm\" (UniqueName: \"kubernetes.io/projected/bcadc39c-6259-49a9-9f5f-38a80545b5a4-kube-api-access-96dcm\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.032655 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.045456 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/556ee9b9-8f10-4de4-9a97-1d3327f885f1-scripts\") pod \"556ee9b9-8f10-4de4-9a97-1d3327f885f1\" (UID: \"556ee9b9-8f10-4de4-9a97-1d3327f885f1\") " Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.045551 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/556ee9b9-8f10-4de4-9a97-1d3327f885f1-httpd-run\") pod \"556ee9b9-8f10-4de4-9a97-1d3327f885f1\" (UID: \"556ee9b9-8f10-4de4-9a97-1d3327f885f1\") " Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.045735 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwvtw\" (UniqueName: \"kubernetes.io/projected/556ee9b9-8f10-4de4-9a97-1d3327f885f1-kube-api-access-hwvtw\") pod \"556ee9b9-8f10-4de4-9a97-1d3327f885f1\" (UID: \"556ee9b9-8f10-4de4-9a97-1d3327f885f1\") " Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.045885 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/556ee9b9-8f10-4de4-9a97-1d3327f885f1-logs\") pod \"556ee9b9-8f10-4de4-9a97-1d3327f885f1\" (UID: \"556ee9b9-8f10-4de4-9a97-1d3327f885f1\") " Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.045929 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"556ee9b9-8f10-4de4-9a97-1d3327f885f1\" (UID: \"556ee9b9-8f10-4de4-9a97-1d3327f885f1\") " Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.045988 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/556ee9b9-8f10-4de4-9a97-1d3327f885f1-config-data\") pod \"556ee9b9-8f10-4de4-9a97-1d3327f885f1\" (UID: \"556ee9b9-8f10-4de4-9a97-1d3327f885f1\") " Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.046039 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/556ee9b9-8f10-4de4-9a97-1d3327f885f1-combined-ca-bundle\") pod \"556ee9b9-8f10-4de4-9a97-1d3327f885f1\" (UID: \"556ee9b9-8f10-4de4-9a97-1d3327f885f1\") " Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.046690 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/556ee9b9-8f10-4de4-9a97-1d3327f885f1-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "556ee9b9-8f10-4de4-9a97-1d3327f885f1" (UID: "556ee9b9-8f10-4de4-9a97-1d3327f885f1"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.046897 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/556ee9b9-8f10-4de4-9a97-1d3327f885f1-logs" (OuterVolumeSpecName: "logs") pod "556ee9b9-8f10-4de4-9a97-1d3327f885f1" (UID: "556ee9b9-8f10-4de4-9a97-1d3327f885f1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.071292 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/556ee9b9-8f10-4de4-9a97-1d3327f885f1-scripts" (OuterVolumeSpecName: "scripts") pod "556ee9b9-8f10-4de4-9a97-1d3327f885f1" (UID: "556ee9b9-8f10-4de4-9a97-1d3327f885f1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.082232 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/556ee9b9-8f10-4de4-9a97-1d3327f885f1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "556ee9b9-8f10-4de4-9a97-1d3327f885f1" (UID: "556ee9b9-8f10-4de4-9a97-1d3327f885f1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.084115 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/556ee9b9-8f10-4de4-9a97-1d3327f885f1-kube-api-access-hwvtw" (OuterVolumeSpecName: "kube-api-access-hwvtw") pod "556ee9b9-8f10-4de4-9a97-1d3327f885f1" (UID: "556ee9b9-8f10-4de4-9a97-1d3327f885f1"). InnerVolumeSpecName "kube-api-access-hwvtw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.084850 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "556ee9b9-8f10-4de4-9a97-1d3327f885f1" (UID: "556ee9b9-8f10-4de4-9a97-1d3327f885f1"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.133795 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/556ee9b9-8f10-4de4-9a97-1d3327f885f1-config-data" (OuterVolumeSpecName: "config-data") pod "556ee9b9-8f10-4de4-9a97-1d3327f885f1" (UID: "556ee9b9-8f10-4de4-9a97-1d3327f885f1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.148065 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwvtw\" (UniqueName: \"kubernetes.io/projected/556ee9b9-8f10-4de4-9a97-1d3327f885f1-kube-api-access-hwvtw\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.148128 4787 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/556ee9b9-8f10-4de4-9a97-1d3327f885f1-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.148164 4787 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.148174 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/556ee9b9-8f10-4de4-9a97-1d3327f885f1-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.148184 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/556ee9b9-8f10-4de4-9a97-1d3327f885f1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.148192 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/556ee9b9-8f10-4de4-9a97-1d3327f885f1-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.148199 4787 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/556ee9b9-8f10-4de4-9a97-1d3327f885f1-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.166891 4787 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.249417 4787 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.397775 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-eafd-account-create-h97jm" event={"ID":"bcadc39c-6259-49a9-9f5f-38a80545b5a4","Type":"ContainerDied","Data":"27779446f41934986444ab73ee658c5bc366fdaef5b2a0c0465ae75308819735"} Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.397845 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27779446f41934986444ab73ee658c5bc366fdaef5b2a0c0465ae75308819735" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.397916 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-eafd-account-create-h97jm" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.409727 4787 generic.go:334] "Generic (PLEG): container finished" podID="556ee9b9-8f10-4de4-9a97-1d3327f885f1" containerID="1c20a8d94201178c836396310098095ed405417f7dcd3ea216a7d70f52076897" exitCode=0 Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.410280 4787 generic.go:334] "Generic (PLEG): container finished" podID="556ee9b9-8f10-4de4-9a97-1d3327f885f1" containerID="9f3da2459ceb85b1e397ec6bdc5977d7ffaadeed925c8538223d712b95ad33c8" exitCode=143 Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.410173 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"556ee9b9-8f10-4de4-9a97-1d3327f885f1","Type":"ContainerDied","Data":"1c20a8d94201178c836396310098095ed405417f7dcd3ea216a7d70f52076897"} Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.410504 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"556ee9b9-8f10-4de4-9a97-1d3327f885f1","Type":"ContainerDied","Data":"9f3da2459ceb85b1e397ec6bdc5977d7ffaadeed925c8538223d712b95ad33c8"} Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.410536 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"556ee9b9-8f10-4de4-9a97-1d3327f885f1","Type":"ContainerDied","Data":"9735648a53d595db56c29fc3e5333ec0d4ec4e25353b8dbce92180cbbefc08bb"} Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.410594 4787 scope.go:117] "RemoveContainer" containerID="1c20a8d94201178c836396310098095ed405417f7dcd3ea216a7d70f52076897" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.410295 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.437023 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3432bc16-99b4-431a-aeb1-600e826bbc3e","Type":"ContainerStarted","Data":"86552a022130c91b0e13f59646e54c5bd18c2d16f8f827243aeb080eaf49dc54"} Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.437189 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3432bc16-99b4-431a-aeb1-600e826bbc3e","Type":"ContainerStarted","Data":"2d4873c7fec8fe1feb92375c09573ee0724d422962f9a29ffb7dfe01ca8981b7"} Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.437264 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3432bc16-99b4-431a-aeb1-600e826bbc3e","Type":"ContainerStarted","Data":"d71dd504167dd253de0aba2a0aab215def36c001f38e2b81df502a9a7884d62a"} Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.521168 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.569139 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.580678 4787 scope.go:117] "RemoveContainer" containerID="9f3da2459ceb85b1e397ec6bdc5977d7ffaadeed925c8538223d712b95ad33c8" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.583773 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:52:27 crc kubenswrapper[4787]: E1001 09:52:27.584192 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22964b11-e82b-4dfa-ac42-70005cbb5caf" containerName="mariadb-account-create" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.584206 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="22964b11-e82b-4dfa-ac42-70005cbb5caf" containerName="mariadb-account-create" Oct 01 09:52:27 crc kubenswrapper[4787]: E1001 09:52:27.584216 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="556ee9b9-8f10-4de4-9a97-1d3327f885f1" containerName="glance-log" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.584223 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="556ee9b9-8f10-4de4-9a97-1d3327f885f1" containerName="glance-log" Oct 01 09:52:27 crc kubenswrapper[4787]: E1001 09:52:27.584234 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="556ee9b9-8f10-4de4-9a97-1d3327f885f1" containerName="glance-httpd" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.584240 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="556ee9b9-8f10-4de4-9a97-1d3327f885f1" containerName="glance-httpd" Oct 01 09:52:27 crc kubenswrapper[4787]: E1001 09:52:27.584265 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcadc39c-6259-49a9-9f5f-38a80545b5a4" containerName="mariadb-account-create" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.584271 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcadc39c-6259-49a9-9f5f-38a80545b5a4" containerName="mariadb-account-create" Oct 01 09:52:27 crc kubenswrapper[4787]: E1001 09:52:27.584290 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54517204-c321-4959-851b-9c642f07a7a7" containerName="mariadb-account-create" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.584296 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="54517204-c321-4959-851b-9c642f07a7a7" containerName="mariadb-account-create" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.584471 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcadc39c-6259-49a9-9f5f-38a80545b5a4" containerName="mariadb-account-create" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.584534 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="54517204-c321-4959-851b-9c642f07a7a7" containerName="mariadb-account-create" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.584551 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="556ee9b9-8f10-4de4-9a97-1d3327f885f1" containerName="glance-log" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.584558 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="22964b11-e82b-4dfa-ac42-70005cbb5caf" containerName="mariadb-account-create" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.584571 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="556ee9b9-8f10-4de4-9a97-1d3327f885f1" containerName="glance-httpd" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.586927 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.591931 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.593002 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.617635 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.662488 4787 scope.go:117] "RemoveContainer" containerID="1c20a8d94201178c836396310098095ed405417f7dcd3ea216a7d70f52076897" Oct 01 09:52:27 crc kubenswrapper[4787]: E1001 09:52:27.663699 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c20a8d94201178c836396310098095ed405417f7dcd3ea216a7d70f52076897\": container with ID starting with 1c20a8d94201178c836396310098095ed405417f7dcd3ea216a7d70f52076897 not found: ID does not exist" containerID="1c20a8d94201178c836396310098095ed405417f7dcd3ea216a7d70f52076897" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.663734 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c20a8d94201178c836396310098095ed405417f7dcd3ea216a7d70f52076897"} err="failed to get container status \"1c20a8d94201178c836396310098095ed405417f7dcd3ea216a7d70f52076897\": rpc error: code = NotFound desc = could not find container \"1c20a8d94201178c836396310098095ed405417f7dcd3ea216a7d70f52076897\": container with ID starting with 1c20a8d94201178c836396310098095ed405417f7dcd3ea216a7d70f52076897 not found: ID does not exist" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.663758 4787 scope.go:117] "RemoveContainer" containerID="9f3da2459ceb85b1e397ec6bdc5977d7ffaadeed925c8538223d712b95ad33c8" Oct 01 09:52:27 crc kubenswrapper[4787]: E1001 09:52:27.664232 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f3da2459ceb85b1e397ec6bdc5977d7ffaadeed925c8538223d712b95ad33c8\": container with ID starting with 9f3da2459ceb85b1e397ec6bdc5977d7ffaadeed925c8538223d712b95ad33c8 not found: ID does not exist" containerID="9f3da2459ceb85b1e397ec6bdc5977d7ffaadeed925c8538223d712b95ad33c8" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.664258 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f3da2459ceb85b1e397ec6bdc5977d7ffaadeed925c8538223d712b95ad33c8"} err="failed to get container status \"9f3da2459ceb85b1e397ec6bdc5977d7ffaadeed925c8538223d712b95ad33c8\": rpc error: code = NotFound desc = could not find container \"9f3da2459ceb85b1e397ec6bdc5977d7ffaadeed925c8538223d712b95ad33c8\": container with ID starting with 9f3da2459ceb85b1e397ec6bdc5977d7ffaadeed925c8538223d712b95ad33c8 not found: ID does not exist" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.664273 4787 scope.go:117] "RemoveContainer" containerID="1c20a8d94201178c836396310098095ed405417f7dcd3ea216a7d70f52076897" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.664798 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c20a8d94201178c836396310098095ed405417f7dcd3ea216a7d70f52076897"} err="failed to get container status \"1c20a8d94201178c836396310098095ed405417f7dcd3ea216a7d70f52076897\": rpc error: code = NotFound desc = could not find container \"1c20a8d94201178c836396310098095ed405417f7dcd3ea216a7d70f52076897\": container with ID starting with 1c20a8d94201178c836396310098095ed405417f7dcd3ea216a7d70f52076897 not found: ID does not exist" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.664822 4787 scope.go:117] "RemoveContainer" containerID="9f3da2459ceb85b1e397ec6bdc5977d7ffaadeed925c8538223d712b95ad33c8" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.665667 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f3da2459ceb85b1e397ec6bdc5977d7ffaadeed925c8538223d712b95ad33c8"} err="failed to get container status \"9f3da2459ceb85b1e397ec6bdc5977d7ffaadeed925c8538223d712b95ad33c8\": rpc error: code = NotFound desc = could not find container \"9f3da2459ceb85b1e397ec6bdc5977d7ffaadeed925c8538223d712b95ad33c8\": container with ID starting with 9f3da2459ceb85b1e397ec6bdc5977d7ffaadeed925c8538223d712b95ad33c8 not found: ID does not exist" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.668969 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.669010 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/527b1b7d-df6a-4d55-8aba-2f1333cd5534-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.669045 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/527b1b7d-df6a-4d55-8aba-2f1333cd5534-config-data\") pod \"glance-default-internal-api-0\" (UID: \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.669089 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/527b1b7d-df6a-4d55-8aba-2f1333cd5534-logs\") pod \"glance-default-internal-api-0\" (UID: \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.669123 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmhwl\" (UniqueName: \"kubernetes.io/projected/527b1b7d-df6a-4d55-8aba-2f1333cd5534-kube-api-access-cmhwl\") pod \"glance-default-internal-api-0\" (UID: \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.669158 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/527b1b7d-df6a-4d55-8aba-2f1333cd5534-scripts\") pod \"glance-default-internal-api-0\" (UID: \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.669215 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/527b1b7d-df6a-4d55-8aba-2f1333cd5534-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.669236 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/527b1b7d-df6a-4d55-8aba-2f1333cd5534-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.773242 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/527b1b7d-df6a-4d55-8aba-2f1333cd5534-logs\") pod \"glance-default-internal-api-0\" (UID: \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.773390 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmhwl\" (UniqueName: \"kubernetes.io/projected/527b1b7d-df6a-4d55-8aba-2f1333cd5534-kube-api-access-cmhwl\") pod \"glance-default-internal-api-0\" (UID: \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.773486 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/527b1b7d-df6a-4d55-8aba-2f1333cd5534-scripts\") pod \"glance-default-internal-api-0\" (UID: \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.773613 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/527b1b7d-df6a-4d55-8aba-2f1333cd5534-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.773653 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/527b1b7d-df6a-4d55-8aba-2f1333cd5534-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.773798 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.773825 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/527b1b7d-df6a-4d55-8aba-2f1333cd5534-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.773879 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/527b1b7d-df6a-4d55-8aba-2f1333cd5534-config-data\") pod \"glance-default-internal-api-0\" (UID: \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.775723 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.777282 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/527b1b7d-df6a-4d55-8aba-2f1333cd5534-logs\") pod \"glance-default-internal-api-0\" (UID: \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.779878 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/527b1b7d-df6a-4d55-8aba-2f1333cd5534-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.791209 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/527b1b7d-df6a-4d55-8aba-2f1333cd5534-scripts\") pod \"glance-default-internal-api-0\" (UID: \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.792267 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/527b1b7d-df6a-4d55-8aba-2f1333cd5534-config-data\") pod \"glance-default-internal-api-0\" (UID: \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.793958 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/527b1b7d-df6a-4d55-8aba-2f1333cd5534-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.794618 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/527b1b7d-df6a-4d55-8aba-2f1333cd5534-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.803521 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmhwl\" (UniqueName: \"kubernetes.io/projected/527b1b7d-df6a-4d55-8aba-2f1333cd5534-kube-api-access-cmhwl\") pod \"glance-default-internal-api-0\" (UID: \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.846722 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.927460 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-slvbz" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.934514 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.977818 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9pxp\" (UniqueName: \"kubernetes.io/projected/6af63266-547b-4537-9290-338b2c0a2d73-kube-api-access-m9pxp\") pod \"6af63266-547b-4537-9290-338b2c0a2d73\" (UID: \"6af63266-547b-4537-9290-338b2c0a2d73\") " Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.977974 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6af63266-547b-4537-9290-338b2c0a2d73-scripts\") pod \"6af63266-547b-4537-9290-338b2c0a2d73\" (UID: \"6af63266-547b-4537-9290-338b2c0a2d73\") " Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.978270 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6af63266-547b-4537-9290-338b2c0a2d73-config-data\") pod \"6af63266-547b-4537-9290-338b2c0a2d73\" (UID: \"6af63266-547b-4537-9290-338b2c0a2d73\") " Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.978299 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6af63266-547b-4537-9290-338b2c0a2d73-combined-ca-bundle\") pod \"6af63266-547b-4537-9290-338b2c0a2d73\" (UID: \"6af63266-547b-4537-9290-338b2c0a2d73\") " Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.978376 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6af63266-547b-4537-9290-338b2c0a2d73-logs\") pod \"6af63266-547b-4537-9290-338b2c0a2d73\" (UID: \"6af63266-547b-4537-9290-338b2c0a2d73\") " Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.981131 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6af63266-547b-4537-9290-338b2c0a2d73-logs" (OuterVolumeSpecName: "logs") pod "6af63266-547b-4537-9290-338b2c0a2d73" (UID: "6af63266-547b-4537-9290-338b2c0a2d73"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.985296 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6af63266-547b-4537-9290-338b2c0a2d73-scripts" (OuterVolumeSpecName: "scripts") pod "6af63266-547b-4537-9290-338b2c0a2d73" (UID: "6af63266-547b-4537-9290-338b2c0a2d73"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:52:27 crc kubenswrapper[4787]: I1001 09:52:27.985309 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6af63266-547b-4537-9290-338b2c0a2d73-kube-api-access-m9pxp" (OuterVolumeSpecName: "kube-api-access-m9pxp") pod "6af63266-547b-4537-9290-338b2c0a2d73" (UID: "6af63266-547b-4537-9290-338b2c0a2d73"). InnerVolumeSpecName "kube-api-access-m9pxp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.019675 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6af63266-547b-4537-9290-338b2c0a2d73-config-data" (OuterVolumeSpecName: "config-data") pod "6af63266-547b-4537-9290-338b2c0a2d73" (UID: "6af63266-547b-4537-9290-338b2c0a2d73"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.034662 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6af63266-547b-4537-9290-338b2c0a2d73-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6af63266-547b-4537-9290-338b2c0a2d73" (UID: "6af63266-547b-4537-9290-338b2c0a2d73"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.083182 4787 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6af63266-547b-4537-9290-338b2c0a2d73-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.083218 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9pxp\" (UniqueName: \"kubernetes.io/projected/6af63266-547b-4537-9290-338b2c0a2d73-kube-api-access-m9pxp\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.083229 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6af63266-547b-4537-9290-338b2c0a2d73-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.083239 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6af63266-547b-4537-9290-338b2c0a2d73-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.083249 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6af63266-547b-4537-9290-338b2c0a2d73-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.471811 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-slvbz" event={"ID":"6af63266-547b-4537-9290-338b2c0a2d73","Type":"ContainerDied","Data":"b8ec9efd63e8d65d90b0fb7729f20b38142a574a75c8a295eb585b2029d2b07c"} Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.471853 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b8ec9efd63e8d65d90b0fb7729f20b38142a574a75c8a295eb585b2029d2b07c" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.471943 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-slvbz" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.605435 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="556ee9b9-8f10-4de4-9a97-1d3327f885f1" path="/var/lib/kubelet/pods/556ee9b9-8f10-4de4-9a97-1d3327f885f1/volumes" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.613063 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3432bc16-99b4-431a-aeb1-600e826bbc3e","Type":"ContainerStarted","Data":"99257767ae580d5ed36236f837206023e0fc53f160d26e7b63489aec240cdabf"} Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.613320 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3432bc16-99b4-431a-aeb1-600e826bbc3e","Type":"ContainerStarted","Data":"79a12a448788bfbdf94500525f799b8e9771393110b13663a76331bc921a6bb6"} Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.613420 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-84d5f4684d-p2xjn"] Oct 01 09:52:28 crc kubenswrapper[4787]: E1001 09:52:28.613862 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6af63266-547b-4537-9290-338b2c0a2d73" containerName="placement-db-sync" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.614170 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="6af63266-547b-4537-9290-338b2c0a2d73" containerName="placement-db-sync" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.616539 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="6af63266-547b-4537-9290-338b2c0a2d73" containerName="placement-db-sync" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.617916 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-84d5f4684d-p2xjn"] Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.619070 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-84d5f4684d-p2xjn" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.628753 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-sfxs7" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.628961 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.629116 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.629243 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.629327 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.713683 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a617751d-ce49-4357-bed0-32a3f63d4d84-config-data\") pod \"placement-84d5f4684d-p2xjn\" (UID: \"a617751d-ce49-4357-bed0-32a3f63d4d84\") " pod="openstack/placement-84d5f4684d-p2xjn" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.713789 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kj48l\" (UniqueName: \"kubernetes.io/projected/a617751d-ce49-4357-bed0-32a3f63d4d84-kube-api-access-kj48l\") pod \"placement-84d5f4684d-p2xjn\" (UID: \"a617751d-ce49-4357-bed0-32a3f63d4d84\") " pod="openstack/placement-84d5f4684d-p2xjn" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.713852 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a617751d-ce49-4357-bed0-32a3f63d4d84-combined-ca-bundle\") pod \"placement-84d5f4684d-p2xjn\" (UID: \"a617751d-ce49-4357-bed0-32a3f63d4d84\") " pod="openstack/placement-84d5f4684d-p2xjn" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.713878 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a617751d-ce49-4357-bed0-32a3f63d4d84-internal-tls-certs\") pod \"placement-84d5f4684d-p2xjn\" (UID: \"a617751d-ce49-4357-bed0-32a3f63d4d84\") " pod="openstack/placement-84d5f4684d-p2xjn" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.713900 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a617751d-ce49-4357-bed0-32a3f63d4d84-scripts\") pod \"placement-84d5f4684d-p2xjn\" (UID: \"a617751d-ce49-4357-bed0-32a3f63d4d84\") " pod="openstack/placement-84d5f4684d-p2xjn" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.713918 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a617751d-ce49-4357-bed0-32a3f63d4d84-public-tls-certs\") pod \"placement-84d5f4684d-p2xjn\" (UID: \"a617751d-ce49-4357-bed0-32a3f63d4d84\") " pod="openstack/placement-84d5f4684d-p2xjn" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.713951 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a617751d-ce49-4357-bed0-32a3f63d4d84-logs\") pod \"placement-84d5f4684d-p2xjn\" (UID: \"a617751d-ce49-4357-bed0-32a3f63d4d84\") " pod="openstack/placement-84d5f4684d-p2xjn" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.796974 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.816457 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a617751d-ce49-4357-bed0-32a3f63d4d84-config-data\") pod \"placement-84d5f4684d-p2xjn\" (UID: \"a617751d-ce49-4357-bed0-32a3f63d4d84\") " pod="openstack/placement-84d5f4684d-p2xjn" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.816541 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kj48l\" (UniqueName: \"kubernetes.io/projected/a617751d-ce49-4357-bed0-32a3f63d4d84-kube-api-access-kj48l\") pod \"placement-84d5f4684d-p2xjn\" (UID: \"a617751d-ce49-4357-bed0-32a3f63d4d84\") " pod="openstack/placement-84d5f4684d-p2xjn" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.816592 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a617751d-ce49-4357-bed0-32a3f63d4d84-combined-ca-bundle\") pod \"placement-84d5f4684d-p2xjn\" (UID: \"a617751d-ce49-4357-bed0-32a3f63d4d84\") " pod="openstack/placement-84d5f4684d-p2xjn" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.816619 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a617751d-ce49-4357-bed0-32a3f63d4d84-internal-tls-certs\") pod \"placement-84d5f4684d-p2xjn\" (UID: \"a617751d-ce49-4357-bed0-32a3f63d4d84\") " pod="openstack/placement-84d5f4684d-p2xjn" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.816643 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a617751d-ce49-4357-bed0-32a3f63d4d84-scripts\") pod \"placement-84d5f4684d-p2xjn\" (UID: \"a617751d-ce49-4357-bed0-32a3f63d4d84\") " pod="openstack/placement-84d5f4684d-p2xjn" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.816660 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a617751d-ce49-4357-bed0-32a3f63d4d84-public-tls-certs\") pod \"placement-84d5f4684d-p2xjn\" (UID: \"a617751d-ce49-4357-bed0-32a3f63d4d84\") " pod="openstack/placement-84d5f4684d-p2xjn" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.816696 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a617751d-ce49-4357-bed0-32a3f63d4d84-logs\") pod \"placement-84d5f4684d-p2xjn\" (UID: \"a617751d-ce49-4357-bed0-32a3f63d4d84\") " pod="openstack/placement-84d5f4684d-p2xjn" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.817113 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a617751d-ce49-4357-bed0-32a3f63d4d84-logs\") pod \"placement-84d5f4684d-p2xjn\" (UID: \"a617751d-ce49-4357-bed0-32a3f63d4d84\") " pod="openstack/placement-84d5f4684d-p2xjn" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.836241 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a617751d-ce49-4357-bed0-32a3f63d4d84-config-data\") pod \"placement-84d5f4684d-p2xjn\" (UID: \"a617751d-ce49-4357-bed0-32a3f63d4d84\") " pod="openstack/placement-84d5f4684d-p2xjn" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.837169 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a617751d-ce49-4357-bed0-32a3f63d4d84-public-tls-certs\") pod \"placement-84d5f4684d-p2xjn\" (UID: \"a617751d-ce49-4357-bed0-32a3f63d4d84\") " pod="openstack/placement-84d5f4684d-p2xjn" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.838462 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a617751d-ce49-4357-bed0-32a3f63d4d84-scripts\") pod \"placement-84d5f4684d-p2xjn\" (UID: \"a617751d-ce49-4357-bed0-32a3f63d4d84\") " pod="openstack/placement-84d5f4684d-p2xjn" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.842416 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a617751d-ce49-4357-bed0-32a3f63d4d84-combined-ca-bundle\") pod \"placement-84d5f4684d-p2xjn\" (UID: \"a617751d-ce49-4357-bed0-32a3f63d4d84\") " pod="openstack/placement-84d5f4684d-p2xjn" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.852148 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a617751d-ce49-4357-bed0-32a3f63d4d84-internal-tls-certs\") pod \"placement-84d5f4684d-p2xjn\" (UID: \"a617751d-ce49-4357-bed0-32a3f63d4d84\") " pod="openstack/placement-84d5f4684d-p2xjn" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.864049 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kj48l\" (UniqueName: \"kubernetes.io/projected/a617751d-ce49-4357-bed0-32a3f63d4d84-kube-api-access-kj48l\") pod \"placement-84d5f4684d-p2xjn\" (UID: \"a617751d-ce49-4357-bed0-32a3f63d4d84\") " pod="openstack/placement-84d5f4684d-p2xjn" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.880391 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-8mrcw"] Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.881841 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-8mrcw" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.884712 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.886279 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-2rt62" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.900917 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.964830 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-8mrcw"] Oct 01 09:52:28 crc kubenswrapper[4787]: I1001 09:52:28.970741 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-84d5f4684d-p2xjn" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.014404 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-wndd6"] Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.015720 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-wndd6" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.020539 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.020716 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-6h4ql" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.023096 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca8b9373-e5e5-4dc1-8995-44ae47abbb13-combined-ca-bundle\") pod \"cinder-db-sync-8mrcw\" (UID: \"ca8b9373-e5e5-4dc1-8995-44ae47abbb13\") " pod="openstack/cinder-db-sync-8mrcw" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.023135 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kfv2\" (UniqueName: \"kubernetes.io/projected/ca8b9373-e5e5-4dc1-8995-44ae47abbb13-kube-api-access-6kfv2\") pod \"cinder-db-sync-8mrcw\" (UID: \"ca8b9373-e5e5-4dc1-8995-44ae47abbb13\") " pod="openstack/cinder-db-sync-8mrcw" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.023188 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ca8b9373-e5e5-4dc1-8995-44ae47abbb13-etc-machine-id\") pod \"cinder-db-sync-8mrcw\" (UID: \"ca8b9373-e5e5-4dc1-8995-44ae47abbb13\") " pod="openstack/cinder-db-sync-8mrcw" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.023207 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ca8b9373-e5e5-4dc1-8995-44ae47abbb13-db-sync-config-data\") pod \"cinder-db-sync-8mrcw\" (UID: \"ca8b9373-e5e5-4dc1-8995-44ae47abbb13\") " pod="openstack/cinder-db-sync-8mrcw" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.023244 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca8b9373-e5e5-4dc1-8995-44ae47abbb13-scripts\") pod \"cinder-db-sync-8mrcw\" (UID: \"ca8b9373-e5e5-4dc1-8995-44ae47abbb13\") " pod="openstack/cinder-db-sync-8mrcw" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.023264 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca8b9373-e5e5-4dc1-8995-44ae47abbb13-config-data\") pod \"cinder-db-sync-8mrcw\" (UID: \"ca8b9373-e5e5-4dc1-8995-44ae47abbb13\") " pod="openstack/cinder-db-sync-8mrcw" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.064671 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-wndd6"] Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.125241 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-kvcbj"] Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.126525 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-kvcbj" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.127210 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15111beb-b05b-440f-a4c9-077eca1c37d1-combined-ca-bundle\") pod \"barbican-db-sync-wndd6\" (UID: \"15111beb-b05b-440f-a4c9-077eca1c37d1\") " pod="openstack/barbican-db-sync-wndd6" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.127275 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca8b9373-e5e5-4dc1-8995-44ae47abbb13-config-data\") pod \"cinder-db-sync-8mrcw\" (UID: \"ca8b9373-e5e5-4dc1-8995-44ae47abbb13\") " pod="openstack/cinder-db-sync-8mrcw" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.127582 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nlgx\" (UniqueName: \"kubernetes.io/projected/15111beb-b05b-440f-a4c9-077eca1c37d1-kube-api-access-8nlgx\") pod \"barbican-db-sync-wndd6\" (UID: \"15111beb-b05b-440f-a4c9-077eca1c37d1\") " pod="openstack/barbican-db-sync-wndd6" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.127631 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca8b9373-e5e5-4dc1-8995-44ae47abbb13-combined-ca-bundle\") pod \"cinder-db-sync-8mrcw\" (UID: \"ca8b9373-e5e5-4dc1-8995-44ae47abbb13\") " pod="openstack/cinder-db-sync-8mrcw" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.127668 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kfv2\" (UniqueName: \"kubernetes.io/projected/ca8b9373-e5e5-4dc1-8995-44ae47abbb13-kube-api-access-6kfv2\") pod \"cinder-db-sync-8mrcw\" (UID: \"ca8b9373-e5e5-4dc1-8995-44ae47abbb13\") " pod="openstack/cinder-db-sync-8mrcw" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.127781 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ca8b9373-e5e5-4dc1-8995-44ae47abbb13-etc-machine-id\") pod \"cinder-db-sync-8mrcw\" (UID: \"ca8b9373-e5e5-4dc1-8995-44ae47abbb13\") " pod="openstack/cinder-db-sync-8mrcw" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.127807 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/15111beb-b05b-440f-a4c9-077eca1c37d1-db-sync-config-data\") pod \"barbican-db-sync-wndd6\" (UID: \"15111beb-b05b-440f-a4c9-077eca1c37d1\") " pod="openstack/barbican-db-sync-wndd6" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.127834 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ca8b9373-e5e5-4dc1-8995-44ae47abbb13-db-sync-config-data\") pod \"cinder-db-sync-8mrcw\" (UID: \"ca8b9373-e5e5-4dc1-8995-44ae47abbb13\") " pod="openstack/cinder-db-sync-8mrcw" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.127946 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca8b9373-e5e5-4dc1-8995-44ae47abbb13-scripts\") pod \"cinder-db-sync-8mrcw\" (UID: \"ca8b9373-e5e5-4dc1-8995-44ae47abbb13\") " pod="openstack/cinder-db-sync-8mrcw" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.133937 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca8b9373-e5e5-4dc1-8995-44ae47abbb13-combined-ca-bundle\") pod \"cinder-db-sync-8mrcw\" (UID: \"ca8b9373-e5e5-4dc1-8995-44ae47abbb13\") " pod="openstack/cinder-db-sync-8mrcw" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.135991 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-jgvxn" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.136334 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.136566 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.140813 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca8b9373-e5e5-4dc1-8995-44ae47abbb13-config-data\") pod \"cinder-db-sync-8mrcw\" (UID: \"ca8b9373-e5e5-4dc1-8995-44ae47abbb13\") " pod="openstack/cinder-db-sync-8mrcw" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.142767 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-kvcbj"] Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.142934 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ca8b9373-e5e5-4dc1-8995-44ae47abbb13-etc-machine-id\") pod \"cinder-db-sync-8mrcw\" (UID: \"ca8b9373-e5e5-4dc1-8995-44ae47abbb13\") " pod="openstack/cinder-db-sync-8mrcw" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.147220 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca8b9373-e5e5-4dc1-8995-44ae47abbb13-scripts\") pod \"cinder-db-sync-8mrcw\" (UID: \"ca8b9373-e5e5-4dc1-8995-44ae47abbb13\") " pod="openstack/cinder-db-sync-8mrcw" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.147424 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ca8b9373-e5e5-4dc1-8995-44ae47abbb13-db-sync-config-data\") pod \"cinder-db-sync-8mrcw\" (UID: \"ca8b9373-e5e5-4dc1-8995-44ae47abbb13\") " pod="openstack/cinder-db-sync-8mrcw" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.178264 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kfv2\" (UniqueName: \"kubernetes.io/projected/ca8b9373-e5e5-4dc1-8995-44ae47abbb13-kube-api-access-6kfv2\") pod \"cinder-db-sync-8mrcw\" (UID: \"ca8b9373-e5e5-4dc1-8995-44ae47abbb13\") " pod="openstack/cinder-db-sync-8mrcw" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.262196 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nlgx\" (UniqueName: \"kubernetes.io/projected/15111beb-b05b-440f-a4c9-077eca1c37d1-kube-api-access-8nlgx\") pod \"barbican-db-sync-wndd6\" (UID: \"15111beb-b05b-440f-a4c9-077eca1c37d1\") " pod="openstack/barbican-db-sync-wndd6" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.262737 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnsq5\" (UniqueName: \"kubernetes.io/projected/18f38986-fd49-4f82-ad99-0a73264877a6-kube-api-access-bnsq5\") pod \"neutron-db-sync-kvcbj\" (UID: \"18f38986-fd49-4f82-ad99-0a73264877a6\") " pod="openstack/neutron-db-sync-kvcbj" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.262874 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/15111beb-b05b-440f-a4c9-077eca1c37d1-db-sync-config-data\") pod \"barbican-db-sync-wndd6\" (UID: \"15111beb-b05b-440f-a4c9-077eca1c37d1\") " pod="openstack/barbican-db-sync-wndd6" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.263024 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15111beb-b05b-440f-a4c9-077eca1c37d1-combined-ca-bundle\") pod \"barbican-db-sync-wndd6\" (UID: \"15111beb-b05b-440f-a4c9-077eca1c37d1\") " pod="openstack/barbican-db-sync-wndd6" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.263102 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18f38986-fd49-4f82-ad99-0a73264877a6-combined-ca-bundle\") pod \"neutron-db-sync-kvcbj\" (UID: \"18f38986-fd49-4f82-ad99-0a73264877a6\") " pod="openstack/neutron-db-sync-kvcbj" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.263152 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/18f38986-fd49-4f82-ad99-0a73264877a6-config\") pod \"neutron-db-sync-kvcbj\" (UID: \"18f38986-fd49-4f82-ad99-0a73264877a6\") " pod="openstack/neutron-db-sync-kvcbj" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.263705 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-8mrcw" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.271296 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15111beb-b05b-440f-a4c9-077eca1c37d1-combined-ca-bundle\") pod \"barbican-db-sync-wndd6\" (UID: \"15111beb-b05b-440f-a4c9-077eca1c37d1\") " pod="openstack/barbican-db-sync-wndd6" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.288854 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/15111beb-b05b-440f-a4c9-077eca1c37d1-db-sync-config-data\") pod \"barbican-db-sync-wndd6\" (UID: \"15111beb-b05b-440f-a4c9-077eca1c37d1\") " pod="openstack/barbican-db-sync-wndd6" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.299697 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nlgx\" (UniqueName: \"kubernetes.io/projected/15111beb-b05b-440f-a4c9-077eca1c37d1-kube-api-access-8nlgx\") pod \"barbican-db-sync-wndd6\" (UID: \"15111beb-b05b-440f-a4c9-077eca1c37d1\") " pod="openstack/barbican-db-sync-wndd6" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.369056 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnsq5\" (UniqueName: \"kubernetes.io/projected/18f38986-fd49-4f82-ad99-0a73264877a6-kube-api-access-bnsq5\") pod \"neutron-db-sync-kvcbj\" (UID: \"18f38986-fd49-4f82-ad99-0a73264877a6\") " pod="openstack/neutron-db-sync-kvcbj" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.369762 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18f38986-fd49-4f82-ad99-0a73264877a6-combined-ca-bundle\") pod \"neutron-db-sync-kvcbj\" (UID: \"18f38986-fd49-4f82-ad99-0a73264877a6\") " pod="openstack/neutron-db-sync-kvcbj" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.369853 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/18f38986-fd49-4f82-ad99-0a73264877a6-config\") pod \"neutron-db-sync-kvcbj\" (UID: \"18f38986-fd49-4f82-ad99-0a73264877a6\") " pod="openstack/neutron-db-sync-kvcbj" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.374337 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/18f38986-fd49-4f82-ad99-0a73264877a6-config\") pod \"neutron-db-sync-kvcbj\" (UID: \"18f38986-fd49-4f82-ad99-0a73264877a6\") " pod="openstack/neutron-db-sync-kvcbj" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.375413 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18f38986-fd49-4f82-ad99-0a73264877a6-combined-ca-bundle\") pod \"neutron-db-sync-kvcbj\" (UID: \"18f38986-fd49-4f82-ad99-0a73264877a6\") " pod="openstack/neutron-db-sync-kvcbj" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.385313 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnsq5\" (UniqueName: \"kubernetes.io/projected/18f38986-fd49-4f82-ad99-0a73264877a6-kube-api-access-bnsq5\") pod \"neutron-db-sync-kvcbj\" (UID: \"18f38986-fd49-4f82-ad99-0a73264877a6\") " pod="openstack/neutron-db-sync-kvcbj" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.571577 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-wndd6" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.601860 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-kvcbj" Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.612615 4787 generic.go:334] "Generic (PLEG): container finished" podID="345ecb87-2ef8-4829-8137-156dec6d0e80" containerID="0657daa702ef6496c3116fb6a7456ba0952ba05d4954c24ac71b5045b3e0c13b" exitCode=0 Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.612712 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dx2gc" event={"ID":"345ecb87-2ef8-4829-8137-156dec6d0e80","Type":"ContainerDied","Data":"0657daa702ef6496c3116fb6a7456ba0952ba05d4954c24ac71b5045b3e0c13b"} Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.630923 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3432bc16-99b4-431a-aeb1-600e826bbc3e","Type":"ContainerStarted","Data":"76a06f9a0655dcc191777411d84113cce08b8a82e8d302e41008df5e89fd8ad8"} Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.646449 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"527b1b7d-df6a-4d55-8aba-2f1333cd5534","Type":"ContainerStarted","Data":"f0029dafde3d93ff6551c565442b5556f4dde57439b6c92dc5ad4146eef37c5c"} Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.659723 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-84d5f4684d-p2xjn"] Oct 01 09:52:29 crc kubenswrapper[4787]: W1001 09:52:29.668262 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda617751d_ce49_4357_bed0_32a3f63d4d84.slice/crio-71c47cf82f4881beb3b2b2fcf9eb3d7958a77cba26114f921ff51dc185d08c6b WatchSource:0}: Error finding container 71c47cf82f4881beb3b2b2fcf9eb3d7958a77cba26114f921ff51dc185d08c6b: Status 404 returned error can't find the container with id 71c47cf82f4881beb3b2b2fcf9eb3d7958a77cba26114f921ff51dc185d08c6b Oct 01 09:52:29 crc kubenswrapper[4787]: I1001 09:52:29.922391 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-8mrcw"] Oct 01 09:52:30 crc kubenswrapper[4787]: I1001 09:52:30.133564 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-wndd6"] Oct 01 09:52:30 crc kubenswrapper[4787]: I1001 09:52:30.268025 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-kvcbj"] Oct 01 09:52:30 crc kubenswrapper[4787]: I1001 09:52:30.659258 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-wndd6" event={"ID":"15111beb-b05b-440f-a4c9-077eca1c37d1","Type":"ContainerStarted","Data":"7d820565c6008eacb9e6bdcd4e423c7a4bb1a7eebd7e955b253647975da4e182"} Oct 01 09:52:30 crc kubenswrapper[4787]: I1001 09:52:30.665138 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-84d5f4684d-p2xjn" event={"ID":"a617751d-ce49-4357-bed0-32a3f63d4d84","Type":"ContainerStarted","Data":"e3034a524c1fad292193b083d78f8ef5ba53336494eebe59dc7657cc072a5ac6"} Oct 01 09:52:30 crc kubenswrapper[4787]: I1001 09:52:30.665187 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-84d5f4684d-p2xjn" event={"ID":"a617751d-ce49-4357-bed0-32a3f63d4d84","Type":"ContainerStarted","Data":"71c47cf82f4881beb3b2b2fcf9eb3d7958a77cba26114f921ff51dc185d08c6b"} Oct 01 09:52:30 crc kubenswrapper[4787]: I1001 09:52:30.668829 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"527b1b7d-df6a-4d55-8aba-2f1333cd5534","Type":"ContainerStarted","Data":"0fe7ee68c24d07552d2277601f96ad684aedcf7f34ffbb273dc8b5ea7a7d2da8"} Oct 01 09:52:30 crc kubenswrapper[4787]: I1001 09:52:30.672306 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-kvcbj" event={"ID":"18f38986-fd49-4f82-ad99-0a73264877a6","Type":"ContainerStarted","Data":"f7622eb300e7d35468fcb5c5811f3de2c4ffb6c7e2df6c995cd08e0572b5abe1"} Oct 01 09:52:30 crc kubenswrapper[4787]: I1001 09:52:30.672371 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-kvcbj" event={"ID":"18f38986-fd49-4f82-ad99-0a73264877a6","Type":"ContainerStarted","Data":"b1b86bd0b9753a43084834f5d143b867460f3e6efc595d650c8a372986af720f"} Oct 01 09:52:30 crc kubenswrapper[4787]: I1001 09:52:30.675827 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-8mrcw" event={"ID":"ca8b9373-e5e5-4dc1-8995-44ae47abbb13","Type":"ContainerStarted","Data":"7dd24c6a97a49814118f912a32a5c7e3c9a96581d698f74375becf0222caa493"} Oct 01 09:52:30 crc kubenswrapper[4787]: I1001 09:52:30.696925 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-kvcbj" podStartSLOduration=1.696899054 podStartE2EDuration="1.696899054s" podCreationTimestamp="2025-10-01 09:52:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:52:30.689598669 +0000 UTC m=+982.804742826" watchObservedRunningTime="2025-10-01 09:52:30.696899054 +0000 UTC m=+982.812043211" Oct 01 09:52:31 crc kubenswrapper[4787]: I1001 09:52:31.086122 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=22.719741001 podStartE2EDuration="1m8.086102299s" podCreationTimestamp="2025-10-01 09:51:23 +0000 UTC" firstStartedPulling="2025-10-01 09:51:41.002946928 +0000 UTC m=+933.118091085" lastFinishedPulling="2025-10-01 09:52:26.369308226 +0000 UTC m=+978.484452383" observedRunningTime="2025-10-01 09:52:30.729623022 +0000 UTC m=+982.844767199" watchObservedRunningTime="2025-10-01 09:52:31.086102299 +0000 UTC m=+983.201246456" Oct 01 09:52:31 crc kubenswrapper[4787]: I1001 09:52:31.091036 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b69d7b58f-t9lc6"] Oct 01 09:52:31 crc kubenswrapper[4787]: I1001 09:52:31.099209 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b69d7b58f-t9lc6" Oct 01 09:52:31 crc kubenswrapper[4787]: I1001 09:52:31.103942 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 01 09:52:31 crc kubenswrapper[4787]: I1001 09:52:31.106192 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b69d7b58f-t9lc6"] Oct 01 09:52:31 crc kubenswrapper[4787]: I1001 09:52:31.219487 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcfg4\" (UniqueName: \"kubernetes.io/projected/6808314c-0e06-4fa4-9c3e-269e5952fd3e-kube-api-access-rcfg4\") pod \"dnsmasq-dns-6b69d7b58f-t9lc6\" (UID: \"6808314c-0e06-4fa4-9c3e-269e5952fd3e\") " pod="openstack/dnsmasq-dns-6b69d7b58f-t9lc6" Oct 01 09:52:31 crc kubenswrapper[4787]: I1001 09:52:31.219570 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6808314c-0e06-4fa4-9c3e-269e5952fd3e-dns-swift-storage-0\") pod \"dnsmasq-dns-6b69d7b58f-t9lc6\" (UID: \"6808314c-0e06-4fa4-9c3e-269e5952fd3e\") " pod="openstack/dnsmasq-dns-6b69d7b58f-t9lc6" Oct 01 09:52:31 crc kubenswrapper[4787]: I1001 09:52:31.219777 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6808314c-0e06-4fa4-9c3e-269e5952fd3e-config\") pod \"dnsmasq-dns-6b69d7b58f-t9lc6\" (UID: \"6808314c-0e06-4fa4-9c3e-269e5952fd3e\") " pod="openstack/dnsmasq-dns-6b69d7b58f-t9lc6" Oct 01 09:52:31 crc kubenswrapper[4787]: I1001 09:52:31.219843 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6808314c-0e06-4fa4-9c3e-269e5952fd3e-ovsdbserver-sb\") pod \"dnsmasq-dns-6b69d7b58f-t9lc6\" (UID: \"6808314c-0e06-4fa4-9c3e-269e5952fd3e\") " pod="openstack/dnsmasq-dns-6b69d7b58f-t9lc6" Oct 01 09:52:31 crc kubenswrapper[4787]: I1001 09:52:31.219877 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6808314c-0e06-4fa4-9c3e-269e5952fd3e-ovsdbserver-nb\") pod \"dnsmasq-dns-6b69d7b58f-t9lc6\" (UID: \"6808314c-0e06-4fa4-9c3e-269e5952fd3e\") " pod="openstack/dnsmasq-dns-6b69d7b58f-t9lc6" Oct 01 09:52:31 crc kubenswrapper[4787]: I1001 09:52:31.220180 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6808314c-0e06-4fa4-9c3e-269e5952fd3e-dns-svc\") pod \"dnsmasq-dns-6b69d7b58f-t9lc6\" (UID: \"6808314c-0e06-4fa4-9c3e-269e5952fd3e\") " pod="openstack/dnsmasq-dns-6b69d7b58f-t9lc6" Oct 01 09:52:31 crc kubenswrapper[4787]: I1001 09:52:31.322300 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6808314c-0e06-4fa4-9c3e-269e5952fd3e-dns-svc\") pod \"dnsmasq-dns-6b69d7b58f-t9lc6\" (UID: \"6808314c-0e06-4fa4-9c3e-269e5952fd3e\") " pod="openstack/dnsmasq-dns-6b69d7b58f-t9lc6" Oct 01 09:52:31 crc kubenswrapper[4787]: I1001 09:52:31.322366 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcfg4\" (UniqueName: \"kubernetes.io/projected/6808314c-0e06-4fa4-9c3e-269e5952fd3e-kube-api-access-rcfg4\") pod \"dnsmasq-dns-6b69d7b58f-t9lc6\" (UID: \"6808314c-0e06-4fa4-9c3e-269e5952fd3e\") " pod="openstack/dnsmasq-dns-6b69d7b58f-t9lc6" Oct 01 09:52:31 crc kubenswrapper[4787]: I1001 09:52:31.322420 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6808314c-0e06-4fa4-9c3e-269e5952fd3e-dns-swift-storage-0\") pod \"dnsmasq-dns-6b69d7b58f-t9lc6\" (UID: \"6808314c-0e06-4fa4-9c3e-269e5952fd3e\") " pod="openstack/dnsmasq-dns-6b69d7b58f-t9lc6" Oct 01 09:52:31 crc kubenswrapper[4787]: I1001 09:52:31.322505 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6808314c-0e06-4fa4-9c3e-269e5952fd3e-config\") pod \"dnsmasq-dns-6b69d7b58f-t9lc6\" (UID: \"6808314c-0e06-4fa4-9c3e-269e5952fd3e\") " pod="openstack/dnsmasq-dns-6b69d7b58f-t9lc6" Oct 01 09:52:31 crc kubenswrapper[4787]: I1001 09:52:31.322524 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6808314c-0e06-4fa4-9c3e-269e5952fd3e-ovsdbserver-sb\") pod \"dnsmasq-dns-6b69d7b58f-t9lc6\" (UID: \"6808314c-0e06-4fa4-9c3e-269e5952fd3e\") " pod="openstack/dnsmasq-dns-6b69d7b58f-t9lc6" Oct 01 09:52:31 crc kubenswrapper[4787]: I1001 09:52:31.322541 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6808314c-0e06-4fa4-9c3e-269e5952fd3e-ovsdbserver-nb\") pod \"dnsmasq-dns-6b69d7b58f-t9lc6\" (UID: \"6808314c-0e06-4fa4-9c3e-269e5952fd3e\") " pod="openstack/dnsmasq-dns-6b69d7b58f-t9lc6" Oct 01 09:52:31 crc kubenswrapper[4787]: I1001 09:52:31.323637 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6808314c-0e06-4fa4-9c3e-269e5952fd3e-dns-swift-storage-0\") pod \"dnsmasq-dns-6b69d7b58f-t9lc6\" (UID: \"6808314c-0e06-4fa4-9c3e-269e5952fd3e\") " pod="openstack/dnsmasq-dns-6b69d7b58f-t9lc6" Oct 01 09:52:31 crc kubenswrapper[4787]: I1001 09:52:31.323730 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6808314c-0e06-4fa4-9c3e-269e5952fd3e-ovsdbserver-nb\") pod \"dnsmasq-dns-6b69d7b58f-t9lc6\" (UID: \"6808314c-0e06-4fa4-9c3e-269e5952fd3e\") " pod="openstack/dnsmasq-dns-6b69d7b58f-t9lc6" Oct 01 09:52:31 crc kubenswrapper[4787]: I1001 09:52:31.323820 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6808314c-0e06-4fa4-9c3e-269e5952fd3e-ovsdbserver-sb\") pod \"dnsmasq-dns-6b69d7b58f-t9lc6\" (UID: \"6808314c-0e06-4fa4-9c3e-269e5952fd3e\") " pod="openstack/dnsmasq-dns-6b69d7b58f-t9lc6" Oct 01 09:52:31 crc kubenswrapper[4787]: I1001 09:52:31.324285 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6808314c-0e06-4fa4-9c3e-269e5952fd3e-dns-svc\") pod \"dnsmasq-dns-6b69d7b58f-t9lc6\" (UID: \"6808314c-0e06-4fa4-9c3e-269e5952fd3e\") " pod="openstack/dnsmasq-dns-6b69d7b58f-t9lc6" Oct 01 09:52:31 crc kubenswrapper[4787]: I1001 09:52:31.324552 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6808314c-0e06-4fa4-9c3e-269e5952fd3e-config\") pod \"dnsmasq-dns-6b69d7b58f-t9lc6\" (UID: \"6808314c-0e06-4fa4-9c3e-269e5952fd3e\") " pod="openstack/dnsmasq-dns-6b69d7b58f-t9lc6" Oct 01 09:52:31 crc kubenswrapper[4787]: I1001 09:52:31.343701 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcfg4\" (UniqueName: \"kubernetes.io/projected/6808314c-0e06-4fa4-9c3e-269e5952fd3e-kube-api-access-rcfg4\") pod \"dnsmasq-dns-6b69d7b58f-t9lc6\" (UID: \"6808314c-0e06-4fa4-9c3e-269e5952fd3e\") " pod="openstack/dnsmasq-dns-6b69d7b58f-t9lc6" Oct 01 09:52:31 crc kubenswrapper[4787]: I1001 09:52:31.431128 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b69d7b58f-t9lc6" Oct 01 09:52:32 crc kubenswrapper[4787]: I1001 09:52:32.599207 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dx2gc" Oct 01 09:52:32 crc kubenswrapper[4787]: I1001 09:52:32.657144 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/345ecb87-2ef8-4829-8137-156dec6d0e80-scripts\") pod \"345ecb87-2ef8-4829-8137-156dec6d0e80\" (UID: \"345ecb87-2ef8-4829-8137-156dec6d0e80\") " Oct 01 09:52:32 crc kubenswrapper[4787]: I1001 09:52:32.657490 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/345ecb87-2ef8-4829-8137-156dec6d0e80-combined-ca-bundle\") pod \"345ecb87-2ef8-4829-8137-156dec6d0e80\" (UID: \"345ecb87-2ef8-4829-8137-156dec6d0e80\") " Oct 01 09:52:32 crc kubenswrapper[4787]: I1001 09:52:32.657658 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/345ecb87-2ef8-4829-8137-156dec6d0e80-config-data\") pod \"345ecb87-2ef8-4829-8137-156dec6d0e80\" (UID: \"345ecb87-2ef8-4829-8137-156dec6d0e80\") " Oct 01 09:52:32 crc kubenswrapper[4787]: I1001 09:52:32.657754 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/345ecb87-2ef8-4829-8137-156dec6d0e80-credential-keys\") pod \"345ecb87-2ef8-4829-8137-156dec6d0e80\" (UID: \"345ecb87-2ef8-4829-8137-156dec6d0e80\") " Oct 01 09:52:32 crc kubenswrapper[4787]: I1001 09:52:32.657873 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/345ecb87-2ef8-4829-8137-156dec6d0e80-fernet-keys\") pod \"345ecb87-2ef8-4829-8137-156dec6d0e80\" (UID: \"345ecb87-2ef8-4829-8137-156dec6d0e80\") " Oct 01 09:52:32 crc kubenswrapper[4787]: I1001 09:52:32.657997 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvfms\" (UniqueName: \"kubernetes.io/projected/345ecb87-2ef8-4829-8137-156dec6d0e80-kube-api-access-jvfms\") pod \"345ecb87-2ef8-4829-8137-156dec6d0e80\" (UID: \"345ecb87-2ef8-4829-8137-156dec6d0e80\") " Oct 01 09:52:32 crc kubenswrapper[4787]: I1001 09:52:32.660756 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/345ecb87-2ef8-4829-8137-156dec6d0e80-scripts" (OuterVolumeSpecName: "scripts") pod "345ecb87-2ef8-4829-8137-156dec6d0e80" (UID: "345ecb87-2ef8-4829-8137-156dec6d0e80"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:52:32 crc kubenswrapper[4787]: I1001 09:52:32.664553 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/345ecb87-2ef8-4829-8137-156dec6d0e80-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "345ecb87-2ef8-4829-8137-156dec6d0e80" (UID: "345ecb87-2ef8-4829-8137-156dec6d0e80"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:52:32 crc kubenswrapper[4787]: I1001 09:52:32.667254 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/345ecb87-2ef8-4829-8137-156dec6d0e80-kube-api-access-jvfms" (OuterVolumeSpecName: "kube-api-access-jvfms") pod "345ecb87-2ef8-4829-8137-156dec6d0e80" (UID: "345ecb87-2ef8-4829-8137-156dec6d0e80"). InnerVolumeSpecName "kube-api-access-jvfms". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:52:32 crc kubenswrapper[4787]: I1001 09:52:32.667622 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/345ecb87-2ef8-4829-8137-156dec6d0e80-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "345ecb87-2ef8-4829-8137-156dec6d0e80" (UID: "345ecb87-2ef8-4829-8137-156dec6d0e80"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:52:32 crc kubenswrapper[4787]: I1001 09:52:32.684334 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/345ecb87-2ef8-4829-8137-156dec6d0e80-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "345ecb87-2ef8-4829-8137-156dec6d0e80" (UID: "345ecb87-2ef8-4829-8137-156dec6d0e80"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:52:32 crc kubenswrapper[4787]: I1001 09:52:32.691761 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/345ecb87-2ef8-4829-8137-156dec6d0e80-config-data" (OuterVolumeSpecName: "config-data") pod "345ecb87-2ef8-4829-8137-156dec6d0e80" (UID: "345ecb87-2ef8-4829-8137-156dec6d0e80"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:52:32 crc kubenswrapper[4787]: I1001 09:52:32.705507 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dx2gc" event={"ID":"345ecb87-2ef8-4829-8137-156dec6d0e80","Type":"ContainerDied","Data":"cee82fc6e04215873beb43900300afa3baed5299bcf23d62d1c87f1deae39f6c"} Oct 01 09:52:32 crc kubenswrapper[4787]: I1001 09:52:32.705583 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cee82fc6e04215873beb43900300afa3baed5299bcf23d62d1c87f1deae39f6c" Oct 01 09:52:32 crc kubenswrapper[4787]: I1001 09:52:32.705794 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dx2gc" Oct 01 09:52:32 crc kubenswrapper[4787]: I1001 09:52:32.760193 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/345ecb87-2ef8-4829-8137-156dec6d0e80-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:32 crc kubenswrapper[4787]: I1001 09:52:32.760232 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/345ecb87-2ef8-4829-8137-156dec6d0e80-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:32 crc kubenswrapper[4787]: I1001 09:52:32.760265 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/345ecb87-2ef8-4829-8137-156dec6d0e80-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:32 crc kubenswrapper[4787]: I1001 09:52:32.760275 4787 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/345ecb87-2ef8-4829-8137-156dec6d0e80-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:32 crc kubenswrapper[4787]: I1001 09:52:32.760286 4787 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/345ecb87-2ef8-4829-8137-156dec6d0e80-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:32 crc kubenswrapper[4787]: I1001 09:52:32.760295 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvfms\" (UniqueName: \"kubernetes.io/projected/345ecb87-2ef8-4829-8137-156dec6d0e80-kube-api-access-jvfms\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:32 crc kubenswrapper[4787]: I1001 09:52:32.883804 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 01 09:52:32 crc kubenswrapper[4787]: I1001 09:52:32.883990 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 01 09:52:32 crc kubenswrapper[4787]: I1001 09:52:32.925795 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 01 09:52:32 crc kubenswrapper[4787]: I1001 09:52:32.930161 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 01 09:52:33 crc kubenswrapper[4787]: I1001 09:52:33.175529 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-67879968b-d5598" Oct 01 09:52:33 crc kubenswrapper[4787]: I1001 09:52:33.176150 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-67879968b-d5598" Oct 01 09:52:33 crc kubenswrapper[4787]: I1001 09:52:33.279619 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7d65d5957b-44rtv" Oct 01 09:52:33 crc kubenswrapper[4787]: I1001 09:52:33.279720 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7d65d5957b-44rtv" Oct 01 09:52:33 crc kubenswrapper[4787]: I1001 09:52:33.714231 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 01 09:52:33 crc kubenswrapper[4787]: I1001 09:52:33.714272 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 01 09:52:33 crc kubenswrapper[4787]: I1001 09:52:33.835153 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-7f57487985-p72bt"] Oct 01 09:52:33 crc kubenswrapper[4787]: E1001 09:52:33.835719 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="345ecb87-2ef8-4829-8137-156dec6d0e80" containerName="keystone-bootstrap" Oct 01 09:52:33 crc kubenswrapper[4787]: I1001 09:52:33.835744 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="345ecb87-2ef8-4829-8137-156dec6d0e80" containerName="keystone-bootstrap" Oct 01 09:52:33 crc kubenswrapper[4787]: I1001 09:52:33.835995 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="345ecb87-2ef8-4829-8137-156dec6d0e80" containerName="keystone-bootstrap" Oct 01 09:52:33 crc kubenswrapper[4787]: I1001 09:52:33.836888 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7f57487985-p72bt" Oct 01 09:52:33 crc kubenswrapper[4787]: I1001 09:52:33.846263 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-qlb5k" Oct 01 09:52:33 crc kubenswrapper[4787]: I1001 09:52:33.846343 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 01 09:52:33 crc kubenswrapper[4787]: I1001 09:52:33.846547 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 01 09:52:33 crc kubenswrapper[4787]: I1001 09:52:33.846610 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 01 09:52:33 crc kubenswrapper[4787]: I1001 09:52:33.846876 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 01 09:52:33 crc kubenswrapper[4787]: I1001 09:52:33.846917 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 01 09:52:33 crc kubenswrapper[4787]: I1001 09:52:33.857339 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7f57487985-p72bt"] Oct 01 09:52:33 crc kubenswrapper[4787]: I1001 09:52:33.883914 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tm57\" (UniqueName: \"kubernetes.io/projected/db5966fa-e6ce-4fd2-9a2c-b17671db2625-kube-api-access-2tm57\") pod \"keystone-7f57487985-p72bt\" (UID: \"db5966fa-e6ce-4fd2-9a2c-b17671db2625\") " pod="openstack/keystone-7f57487985-p72bt" Oct 01 09:52:33 crc kubenswrapper[4787]: I1001 09:52:33.884002 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/db5966fa-e6ce-4fd2-9a2c-b17671db2625-fernet-keys\") pod \"keystone-7f57487985-p72bt\" (UID: \"db5966fa-e6ce-4fd2-9a2c-b17671db2625\") " pod="openstack/keystone-7f57487985-p72bt" Oct 01 09:52:33 crc kubenswrapper[4787]: I1001 09:52:33.884040 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db5966fa-e6ce-4fd2-9a2c-b17671db2625-combined-ca-bundle\") pod \"keystone-7f57487985-p72bt\" (UID: \"db5966fa-e6ce-4fd2-9a2c-b17671db2625\") " pod="openstack/keystone-7f57487985-p72bt" Oct 01 09:52:33 crc kubenswrapper[4787]: I1001 09:52:33.884095 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/db5966fa-e6ce-4fd2-9a2c-b17671db2625-internal-tls-certs\") pod \"keystone-7f57487985-p72bt\" (UID: \"db5966fa-e6ce-4fd2-9a2c-b17671db2625\") " pod="openstack/keystone-7f57487985-p72bt" Oct 01 09:52:33 crc kubenswrapper[4787]: I1001 09:52:33.884121 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/db5966fa-e6ce-4fd2-9a2c-b17671db2625-public-tls-certs\") pod \"keystone-7f57487985-p72bt\" (UID: \"db5966fa-e6ce-4fd2-9a2c-b17671db2625\") " pod="openstack/keystone-7f57487985-p72bt" Oct 01 09:52:33 crc kubenswrapper[4787]: I1001 09:52:33.884141 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db5966fa-e6ce-4fd2-9a2c-b17671db2625-scripts\") pod \"keystone-7f57487985-p72bt\" (UID: \"db5966fa-e6ce-4fd2-9a2c-b17671db2625\") " pod="openstack/keystone-7f57487985-p72bt" Oct 01 09:52:33 crc kubenswrapper[4787]: I1001 09:52:33.884180 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/db5966fa-e6ce-4fd2-9a2c-b17671db2625-credential-keys\") pod \"keystone-7f57487985-p72bt\" (UID: \"db5966fa-e6ce-4fd2-9a2c-b17671db2625\") " pod="openstack/keystone-7f57487985-p72bt" Oct 01 09:52:33 crc kubenswrapper[4787]: I1001 09:52:33.884213 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db5966fa-e6ce-4fd2-9a2c-b17671db2625-config-data\") pod \"keystone-7f57487985-p72bt\" (UID: \"db5966fa-e6ce-4fd2-9a2c-b17671db2625\") " pod="openstack/keystone-7f57487985-p72bt" Oct 01 09:52:33 crc kubenswrapper[4787]: I1001 09:52:33.985268 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/db5966fa-e6ce-4fd2-9a2c-b17671db2625-credential-keys\") pod \"keystone-7f57487985-p72bt\" (UID: \"db5966fa-e6ce-4fd2-9a2c-b17671db2625\") " pod="openstack/keystone-7f57487985-p72bt" Oct 01 09:52:33 crc kubenswrapper[4787]: I1001 09:52:33.985332 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db5966fa-e6ce-4fd2-9a2c-b17671db2625-config-data\") pod \"keystone-7f57487985-p72bt\" (UID: \"db5966fa-e6ce-4fd2-9a2c-b17671db2625\") " pod="openstack/keystone-7f57487985-p72bt" Oct 01 09:52:33 crc kubenswrapper[4787]: I1001 09:52:33.985385 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tm57\" (UniqueName: \"kubernetes.io/projected/db5966fa-e6ce-4fd2-9a2c-b17671db2625-kube-api-access-2tm57\") pod \"keystone-7f57487985-p72bt\" (UID: \"db5966fa-e6ce-4fd2-9a2c-b17671db2625\") " pod="openstack/keystone-7f57487985-p72bt" Oct 01 09:52:33 crc kubenswrapper[4787]: I1001 09:52:33.985413 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/db5966fa-e6ce-4fd2-9a2c-b17671db2625-fernet-keys\") pod \"keystone-7f57487985-p72bt\" (UID: \"db5966fa-e6ce-4fd2-9a2c-b17671db2625\") " pod="openstack/keystone-7f57487985-p72bt" Oct 01 09:52:33 crc kubenswrapper[4787]: I1001 09:52:33.985450 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db5966fa-e6ce-4fd2-9a2c-b17671db2625-combined-ca-bundle\") pod \"keystone-7f57487985-p72bt\" (UID: \"db5966fa-e6ce-4fd2-9a2c-b17671db2625\") " pod="openstack/keystone-7f57487985-p72bt" Oct 01 09:52:33 crc kubenswrapper[4787]: I1001 09:52:33.985490 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/db5966fa-e6ce-4fd2-9a2c-b17671db2625-internal-tls-certs\") pod \"keystone-7f57487985-p72bt\" (UID: \"db5966fa-e6ce-4fd2-9a2c-b17671db2625\") " pod="openstack/keystone-7f57487985-p72bt" Oct 01 09:52:33 crc kubenswrapper[4787]: I1001 09:52:33.985511 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/db5966fa-e6ce-4fd2-9a2c-b17671db2625-public-tls-certs\") pod \"keystone-7f57487985-p72bt\" (UID: \"db5966fa-e6ce-4fd2-9a2c-b17671db2625\") " pod="openstack/keystone-7f57487985-p72bt" Oct 01 09:52:33 crc kubenswrapper[4787]: I1001 09:52:33.985534 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db5966fa-e6ce-4fd2-9a2c-b17671db2625-scripts\") pod \"keystone-7f57487985-p72bt\" (UID: \"db5966fa-e6ce-4fd2-9a2c-b17671db2625\") " pod="openstack/keystone-7f57487985-p72bt" Oct 01 09:52:33 crc kubenswrapper[4787]: I1001 09:52:33.992156 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/db5966fa-e6ce-4fd2-9a2c-b17671db2625-public-tls-certs\") pod \"keystone-7f57487985-p72bt\" (UID: \"db5966fa-e6ce-4fd2-9a2c-b17671db2625\") " pod="openstack/keystone-7f57487985-p72bt" Oct 01 09:52:33 crc kubenswrapper[4787]: I1001 09:52:33.995735 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db5966fa-e6ce-4fd2-9a2c-b17671db2625-scripts\") pod \"keystone-7f57487985-p72bt\" (UID: \"db5966fa-e6ce-4fd2-9a2c-b17671db2625\") " pod="openstack/keystone-7f57487985-p72bt" Oct 01 09:52:33 crc kubenswrapper[4787]: I1001 09:52:33.996291 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/db5966fa-e6ce-4fd2-9a2c-b17671db2625-credential-keys\") pod \"keystone-7f57487985-p72bt\" (UID: \"db5966fa-e6ce-4fd2-9a2c-b17671db2625\") " pod="openstack/keystone-7f57487985-p72bt" Oct 01 09:52:33 crc kubenswrapper[4787]: I1001 09:52:33.996561 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/db5966fa-e6ce-4fd2-9a2c-b17671db2625-fernet-keys\") pod \"keystone-7f57487985-p72bt\" (UID: \"db5966fa-e6ce-4fd2-9a2c-b17671db2625\") " pod="openstack/keystone-7f57487985-p72bt" Oct 01 09:52:33 crc kubenswrapper[4787]: I1001 09:52:33.996920 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db5966fa-e6ce-4fd2-9a2c-b17671db2625-config-data\") pod \"keystone-7f57487985-p72bt\" (UID: \"db5966fa-e6ce-4fd2-9a2c-b17671db2625\") " pod="openstack/keystone-7f57487985-p72bt" Oct 01 09:52:33 crc kubenswrapper[4787]: I1001 09:52:33.997246 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db5966fa-e6ce-4fd2-9a2c-b17671db2625-combined-ca-bundle\") pod \"keystone-7f57487985-p72bt\" (UID: \"db5966fa-e6ce-4fd2-9a2c-b17671db2625\") " pod="openstack/keystone-7f57487985-p72bt" Oct 01 09:52:33 crc kubenswrapper[4787]: I1001 09:52:33.997495 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/db5966fa-e6ce-4fd2-9a2c-b17671db2625-internal-tls-certs\") pod \"keystone-7f57487985-p72bt\" (UID: \"db5966fa-e6ce-4fd2-9a2c-b17671db2625\") " pod="openstack/keystone-7f57487985-p72bt" Oct 01 09:52:34 crc kubenswrapper[4787]: I1001 09:52:34.006140 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tm57\" (UniqueName: \"kubernetes.io/projected/db5966fa-e6ce-4fd2-9a2c-b17671db2625-kube-api-access-2tm57\") pod \"keystone-7f57487985-p72bt\" (UID: \"db5966fa-e6ce-4fd2-9a2c-b17671db2625\") " pod="openstack/keystone-7f57487985-p72bt" Oct 01 09:52:34 crc kubenswrapper[4787]: I1001 09:52:34.160400 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7f57487985-p72bt" Oct 01 09:52:35 crc kubenswrapper[4787]: I1001 09:52:35.217686 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b69d7b58f-t9lc6"] Oct 01 09:52:35 crc kubenswrapper[4787]: I1001 09:52:35.309890 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7f57487985-p72bt"] Oct 01 09:52:35 crc kubenswrapper[4787]: I1001 09:52:35.764383 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9587d64e-33e3-45d2-8ee7-a776fcc60d88","Type":"ContainerStarted","Data":"8b380d576833d002f4a4c98eff2e55905be1be5916394c71d0684713baac85ef"} Oct 01 09:52:35 crc kubenswrapper[4787]: I1001 09:52:35.772326 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"527b1b7d-df6a-4d55-8aba-2f1333cd5534","Type":"ContainerStarted","Data":"0875d6598ced04ae5adf41e252c6f3988626ac39ce64ccae48a815ad87ee8791"} Oct 01 09:52:35 crc kubenswrapper[4787]: I1001 09:52:35.784978 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7f57487985-p72bt" event={"ID":"db5966fa-e6ce-4fd2-9a2c-b17671db2625","Type":"ContainerStarted","Data":"858976640ae5c250d5badb38b2bd02ebe3824479788ae1368400197c17b34a84"} Oct 01 09:52:35 crc kubenswrapper[4787]: I1001 09:52:35.785049 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7f57487985-p72bt" event={"ID":"db5966fa-e6ce-4fd2-9a2c-b17671db2625","Type":"ContainerStarted","Data":"db35fc8597db93f371553eabf34cb4839b2225458519492c34ebbe23d7b7e2d1"} Oct 01 09:52:35 crc kubenswrapper[4787]: I1001 09:52:35.785116 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-7f57487985-p72bt" Oct 01 09:52:35 crc kubenswrapper[4787]: I1001 09:52:35.790929 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-84d5f4684d-p2xjn" event={"ID":"a617751d-ce49-4357-bed0-32a3f63d4d84","Type":"ContainerStarted","Data":"4f08947c2a5a9089b4eab82462f65e9deeb404b3666017d725687958949a2ee0"} Oct 01 09:52:35 crc kubenswrapper[4787]: I1001 09:52:35.791177 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-84d5f4684d-p2xjn" Oct 01 09:52:35 crc kubenswrapper[4787]: I1001 09:52:35.792250 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-84d5f4684d-p2xjn" Oct 01 09:52:35 crc kubenswrapper[4787]: I1001 09:52:35.799315 4787 generic.go:334] "Generic (PLEG): container finished" podID="6808314c-0e06-4fa4-9c3e-269e5952fd3e" containerID="7871e8dfa99a35696ecc214ea9f6b3f9e404c4b4ca98c24c34741b14a1fb92a1" exitCode=0 Oct 01 09:52:35 crc kubenswrapper[4787]: I1001 09:52:35.799434 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b69d7b58f-t9lc6" event={"ID":"6808314c-0e06-4fa4-9c3e-269e5952fd3e","Type":"ContainerDied","Data":"7871e8dfa99a35696ecc214ea9f6b3f9e404c4b4ca98c24c34741b14a1fb92a1"} Oct 01 09:52:35 crc kubenswrapper[4787]: I1001 09:52:35.799463 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b69d7b58f-t9lc6" event={"ID":"6808314c-0e06-4fa4-9c3e-269e5952fd3e","Type":"ContainerStarted","Data":"199c7ef098dbbfd089ac3178bd69c016445a7ec524defee9b244ef53ba6497b7"} Oct 01 09:52:35 crc kubenswrapper[4787]: I1001 09:52:35.800925 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=8.80089057 podStartE2EDuration="8.80089057s" podCreationTimestamp="2025-10-01 09:52:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:52:35.790865707 +0000 UTC m=+987.906009864" watchObservedRunningTime="2025-10-01 09:52:35.80089057 +0000 UTC m=+987.916034727" Oct 01 09:52:35 crc kubenswrapper[4787]: I1001 09:52:35.820671 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-7f57487985-p72bt" podStartSLOduration=2.820648231 podStartE2EDuration="2.820648231s" podCreationTimestamp="2025-10-01 09:52:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:52:35.816210328 +0000 UTC m=+987.931354495" watchObservedRunningTime="2025-10-01 09:52:35.820648231 +0000 UTC m=+987.935792388" Oct 01 09:52:35 crc kubenswrapper[4787]: I1001 09:52:35.888510 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-84d5f4684d-p2xjn" podStartSLOduration=7.888490169 podStartE2EDuration="7.888490169s" podCreationTimestamp="2025-10-01 09:52:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:52:35.878163187 +0000 UTC m=+987.993307364" watchObservedRunningTime="2025-10-01 09:52:35.888490169 +0000 UTC m=+988.003634326" Oct 01 09:52:36 crc kubenswrapper[4787]: I1001 09:52:36.077382 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 01 09:52:36 crc kubenswrapper[4787]: I1001 09:52:36.077480 4787 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 01 09:52:36 crc kubenswrapper[4787]: I1001 09:52:36.541154 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 01 09:52:36 crc kubenswrapper[4787]: I1001 09:52:36.821554 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b69d7b58f-t9lc6" event={"ID":"6808314c-0e06-4fa4-9c3e-269e5952fd3e","Type":"ContainerStarted","Data":"25a024f9e9cad11ef9f8ce3dd6a938d6dba2bec6ef4e26686340150ab47e10f9"} Oct 01 09:52:36 crc kubenswrapper[4787]: I1001 09:52:36.822763 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b69d7b58f-t9lc6" Oct 01 09:52:36 crc kubenswrapper[4787]: I1001 09:52:36.853204 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b69d7b58f-t9lc6" podStartSLOduration=5.853186276 podStartE2EDuration="5.853186276s" podCreationTimestamp="2025-10-01 09:52:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:52:36.852300303 +0000 UTC m=+988.967444460" watchObservedRunningTime="2025-10-01 09:52:36.853186276 +0000 UTC m=+988.968330433" Oct 01 09:52:37 crc kubenswrapper[4787]: I1001 09:52:37.178694 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-84d5f4684d-p2xjn" Oct 01 09:52:37 crc kubenswrapper[4787]: I1001 09:52:37.935181 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 01 09:52:37 crc kubenswrapper[4787]: I1001 09:52:37.935606 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 01 09:52:37 crc kubenswrapper[4787]: I1001 09:52:37.978382 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 01 09:52:37 crc kubenswrapper[4787]: I1001 09:52:37.997582 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 01 09:52:38 crc kubenswrapper[4787]: I1001 09:52:38.852526 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 01 09:52:38 crc kubenswrapper[4787]: I1001 09:52:38.852582 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 01 09:52:40 crc kubenswrapper[4787]: I1001 09:52:40.872888 4787 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 01 09:52:40 crc kubenswrapper[4787]: I1001 09:52:40.999400 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 01 09:52:41 crc kubenswrapper[4787]: I1001 09:52:41.001264 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 01 09:52:41 crc kubenswrapper[4787]: I1001 09:52:41.436193 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b69d7b58f-t9lc6" Oct 01 09:52:41 crc kubenswrapper[4787]: I1001 09:52:41.528379 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bb97fccb5-vblx7"] Oct 01 09:52:41 crc kubenswrapper[4787]: I1001 09:52:41.528658 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7bb97fccb5-vblx7" podUID="a63d659d-52a4-4f33-85c2-1461423776a3" containerName="dnsmasq-dns" containerID="cri-o://4e7e606d8da369890eb621b428c3ac2e5da4b3dfcc45f7505cf0efc6d81bfb76" gracePeriod=10 Oct 01 09:52:41 crc kubenswrapper[4787]: I1001 09:52:41.696730 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7bb97fccb5-vblx7" podUID="a63d659d-52a4-4f33-85c2-1461423776a3" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.138:5353: connect: connection refused" Oct 01 09:52:41 crc kubenswrapper[4787]: I1001 09:52:41.889460 4787 generic.go:334] "Generic (PLEG): container finished" podID="a63d659d-52a4-4f33-85c2-1461423776a3" containerID="4e7e606d8da369890eb621b428c3ac2e5da4b3dfcc45f7505cf0efc6d81bfb76" exitCode=0 Oct 01 09:52:41 crc kubenswrapper[4787]: I1001 09:52:41.889805 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bb97fccb5-vblx7" event={"ID":"a63d659d-52a4-4f33-85c2-1461423776a3","Type":"ContainerDied","Data":"4e7e606d8da369890eb621b428c3ac2e5da4b3dfcc45f7505cf0efc6d81bfb76"} Oct 01 09:52:43 crc kubenswrapper[4787]: I1001 09:52:43.180098 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-67879968b-d5598" podUID="7983e213-c317-42b2-8944-85bdb78dac8f" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Oct 01 09:52:43 crc kubenswrapper[4787]: I1001 09:52:43.282931 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7d65d5957b-44rtv" podUID="db1d14ba-8d00-4508-9c89-fdadb562ad0d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Oct 01 09:52:49 crc kubenswrapper[4787]: I1001 09:52:49.993792 4787 generic.go:334] "Generic (PLEG): container finished" podID="18f38986-fd49-4f82-ad99-0a73264877a6" containerID="f7622eb300e7d35468fcb5c5811f3de2c4ffb6c7e2df6c995cd08e0572b5abe1" exitCode=0 Oct 01 09:52:49 crc kubenswrapper[4787]: I1001 09:52:49.993872 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-kvcbj" event={"ID":"18f38986-fd49-4f82-ad99-0a73264877a6","Type":"ContainerDied","Data":"f7622eb300e7d35468fcb5c5811f3de2c4ffb6c7e2df6c995cd08e0572b5abe1"} Oct 01 09:52:51 crc kubenswrapper[4787]: I1001 09:52:51.696147 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7bb97fccb5-vblx7" podUID="a63d659d-52a4-4f33-85c2-1461423776a3" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.138:5353: i/o timeout" Oct 01 09:52:54 crc kubenswrapper[4787]: I1001 09:52:54.209827 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bb97fccb5-vblx7" Oct 01 09:52:54 crc kubenswrapper[4787]: I1001 09:52:54.226445 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljbw8\" (UniqueName: \"kubernetes.io/projected/a63d659d-52a4-4f33-85c2-1461423776a3-kube-api-access-ljbw8\") pod \"a63d659d-52a4-4f33-85c2-1461423776a3\" (UID: \"a63d659d-52a4-4f33-85c2-1461423776a3\") " Oct 01 09:52:54 crc kubenswrapper[4787]: I1001 09:52:54.226972 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a63d659d-52a4-4f33-85c2-1461423776a3-config\") pod \"a63d659d-52a4-4f33-85c2-1461423776a3\" (UID: \"a63d659d-52a4-4f33-85c2-1461423776a3\") " Oct 01 09:52:54 crc kubenswrapper[4787]: I1001 09:52:54.227007 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a63d659d-52a4-4f33-85c2-1461423776a3-ovsdbserver-sb\") pod \"a63d659d-52a4-4f33-85c2-1461423776a3\" (UID: \"a63d659d-52a4-4f33-85c2-1461423776a3\") " Oct 01 09:52:54 crc kubenswrapper[4787]: I1001 09:52:54.234110 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a63d659d-52a4-4f33-85c2-1461423776a3-kube-api-access-ljbw8" (OuterVolumeSpecName: "kube-api-access-ljbw8") pod "a63d659d-52a4-4f33-85c2-1461423776a3" (UID: "a63d659d-52a4-4f33-85c2-1461423776a3"). InnerVolumeSpecName "kube-api-access-ljbw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:52:54 crc kubenswrapper[4787]: I1001 09:52:54.276384 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a63d659d-52a4-4f33-85c2-1461423776a3-config" (OuterVolumeSpecName: "config") pod "a63d659d-52a4-4f33-85c2-1461423776a3" (UID: "a63d659d-52a4-4f33-85c2-1461423776a3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:52:54 crc kubenswrapper[4787]: I1001 09:52:54.294050 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a63d659d-52a4-4f33-85c2-1461423776a3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a63d659d-52a4-4f33-85c2-1461423776a3" (UID: "a63d659d-52a4-4f33-85c2-1461423776a3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:52:54 crc kubenswrapper[4787]: I1001 09:52:54.327982 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a63d659d-52a4-4f33-85c2-1461423776a3-dns-svc\") pod \"a63d659d-52a4-4f33-85c2-1461423776a3\" (UID: \"a63d659d-52a4-4f33-85c2-1461423776a3\") " Oct 01 09:52:54 crc kubenswrapper[4787]: I1001 09:52:54.328041 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a63d659d-52a4-4f33-85c2-1461423776a3-ovsdbserver-nb\") pod \"a63d659d-52a4-4f33-85c2-1461423776a3\" (UID: \"a63d659d-52a4-4f33-85c2-1461423776a3\") " Oct 01 09:52:54 crc kubenswrapper[4787]: I1001 09:52:54.328305 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljbw8\" (UniqueName: \"kubernetes.io/projected/a63d659d-52a4-4f33-85c2-1461423776a3-kube-api-access-ljbw8\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:54 crc kubenswrapper[4787]: I1001 09:52:54.328340 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a63d659d-52a4-4f33-85c2-1461423776a3-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:54 crc kubenswrapper[4787]: I1001 09:52:54.328349 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a63d659d-52a4-4f33-85c2-1461423776a3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:54 crc kubenswrapper[4787]: I1001 09:52:54.371793 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a63d659d-52a4-4f33-85c2-1461423776a3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a63d659d-52a4-4f33-85c2-1461423776a3" (UID: "a63d659d-52a4-4f33-85c2-1461423776a3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:52:54 crc kubenswrapper[4787]: I1001 09:52:54.381040 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a63d659d-52a4-4f33-85c2-1461423776a3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a63d659d-52a4-4f33-85c2-1461423776a3" (UID: "a63d659d-52a4-4f33-85c2-1461423776a3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:52:54 crc kubenswrapper[4787]: I1001 09:52:54.429946 4787 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a63d659d-52a4-4f33-85c2-1461423776a3-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:54 crc kubenswrapper[4787]: I1001 09:52:54.429987 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a63d659d-52a4-4f33-85c2-1461423776a3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:54 crc kubenswrapper[4787]: I1001 09:52:54.993922 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-67879968b-d5598" Oct 01 09:52:55 crc kubenswrapper[4787]: I1001 09:52:55.040492 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bb97fccb5-vblx7" event={"ID":"a63d659d-52a4-4f33-85c2-1461423776a3","Type":"ContainerDied","Data":"acfe6f4935f074301a7a9f7a2bd89dfb417aa356089cdb784bda91e6255a9104"} Oct 01 09:52:55 crc kubenswrapper[4787]: I1001 09:52:55.040555 4787 scope.go:117] "RemoveContainer" containerID="4e7e606d8da369890eb621b428c3ac2e5da4b3dfcc45f7505cf0efc6d81bfb76" Oct 01 09:52:55 crc kubenswrapper[4787]: I1001 09:52:55.040713 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bb97fccb5-vblx7" Oct 01 09:52:55 crc kubenswrapper[4787]: I1001 09:52:55.067444 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bb97fccb5-vblx7"] Oct 01 09:52:55 crc kubenswrapper[4787]: I1001 09:52:55.073945 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7bb97fccb5-vblx7"] Oct 01 09:52:55 crc kubenswrapper[4787]: I1001 09:52:55.167712 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7d65d5957b-44rtv" Oct 01 09:52:55 crc kubenswrapper[4787]: E1001 09:52:55.606981 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:e318869f706836a0c74c0ad55aab277b1bb7fae0555ae0f03cb28b379b9ce695" Oct 01 09:52:55 crc kubenswrapper[4787]: E1001 09:52:55.607208 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:e318869f706836a0c74c0ad55aab277b1bb7fae0555ae0f03cb28b379b9ce695,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6kfv2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-8mrcw_openstack(ca8b9373-e5e5-4dc1-8995-44ae47abbb13): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 09:52:55 crc kubenswrapper[4787]: E1001 09:52:55.608708 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-8mrcw" podUID="ca8b9373-e5e5-4dc1-8995-44ae47abbb13" Oct 01 09:52:55 crc kubenswrapper[4787]: I1001 09:52:55.640254 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-kvcbj" Oct 01 09:52:55 crc kubenswrapper[4787]: I1001 09:52:55.672405 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18f38986-fd49-4f82-ad99-0a73264877a6-combined-ca-bundle\") pod \"18f38986-fd49-4f82-ad99-0a73264877a6\" (UID: \"18f38986-fd49-4f82-ad99-0a73264877a6\") " Oct 01 09:52:55 crc kubenswrapper[4787]: I1001 09:52:55.672649 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnsq5\" (UniqueName: \"kubernetes.io/projected/18f38986-fd49-4f82-ad99-0a73264877a6-kube-api-access-bnsq5\") pod \"18f38986-fd49-4f82-ad99-0a73264877a6\" (UID: \"18f38986-fd49-4f82-ad99-0a73264877a6\") " Oct 01 09:52:55 crc kubenswrapper[4787]: I1001 09:52:55.673611 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/18f38986-fd49-4f82-ad99-0a73264877a6-config\") pod \"18f38986-fd49-4f82-ad99-0a73264877a6\" (UID: \"18f38986-fd49-4f82-ad99-0a73264877a6\") " Oct 01 09:52:55 crc kubenswrapper[4787]: I1001 09:52:55.681113 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18f38986-fd49-4f82-ad99-0a73264877a6-kube-api-access-bnsq5" (OuterVolumeSpecName: "kube-api-access-bnsq5") pod "18f38986-fd49-4f82-ad99-0a73264877a6" (UID: "18f38986-fd49-4f82-ad99-0a73264877a6"). InnerVolumeSpecName "kube-api-access-bnsq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:52:55 crc kubenswrapper[4787]: I1001 09:52:55.709643 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18f38986-fd49-4f82-ad99-0a73264877a6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "18f38986-fd49-4f82-ad99-0a73264877a6" (UID: "18f38986-fd49-4f82-ad99-0a73264877a6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:52:55 crc kubenswrapper[4787]: I1001 09:52:55.725827 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18f38986-fd49-4f82-ad99-0a73264877a6-config" (OuterVolumeSpecName: "config") pod "18f38986-fd49-4f82-ad99-0a73264877a6" (UID: "18f38986-fd49-4f82-ad99-0a73264877a6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:52:55 crc kubenswrapper[4787]: I1001 09:52:55.775043 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnsq5\" (UniqueName: \"kubernetes.io/projected/18f38986-fd49-4f82-ad99-0a73264877a6-kube-api-access-bnsq5\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:55 crc kubenswrapper[4787]: I1001 09:52:55.775118 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/18f38986-fd49-4f82-ad99-0a73264877a6-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:55 crc kubenswrapper[4787]: I1001 09:52:55.775136 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18f38986-fd49-4f82-ad99-0a73264877a6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:56 crc kubenswrapper[4787]: I1001 09:52:56.055499 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-kvcbj" Oct 01 09:52:56 crc kubenswrapper[4787]: I1001 09:52:56.055520 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-kvcbj" event={"ID":"18f38986-fd49-4f82-ad99-0a73264877a6","Type":"ContainerDied","Data":"b1b86bd0b9753a43084834f5d143b867460f3e6efc595d650c8a372986af720f"} Oct 01 09:52:56 crc kubenswrapper[4787]: I1001 09:52:56.055586 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1b86bd0b9753a43084834f5d143b867460f3e6efc595d650c8a372986af720f" Oct 01 09:52:56 crc kubenswrapper[4787]: I1001 09:52:56.080176 4787 scope.go:117] "RemoveContainer" containerID="b4b7dcf91d6035d9dc19d6070186f3ddafa3f9f0b196a665c5a018cd70250c71" Oct 01 09:52:56 crc kubenswrapper[4787]: E1001 09:52:56.080431 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:e318869f706836a0c74c0ad55aab277b1bb7fae0555ae0f03cb28b379b9ce695\\\"\"" pod="openstack/cinder-db-sync-8mrcw" podUID="ca8b9373-e5e5-4dc1-8995-44ae47abbb13" Oct 01 09:52:56 crc kubenswrapper[4787]: E1001 09:52:56.083004 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/ubi9/httpd-24@sha256:e91d58021b54c46883595ff66be65882de54abdb3be2ca53c4162b20d18b5f48" Oct 01 09:52:56 crc kubenswrapper[4787]: E1001 09:52:56.083292 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:proxy-httpd,Image:registry.redhat.io/ubi9/httpd-24@sha256:e91d58021b54c46883595ff66be65882de54abdb3be2ca53c4162b20d18b5f48,Command:[/usr/sbin/httpd],Args:[-DFOREGROUND],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:proxy-httpd,HostPort:0,ContainerPort:3000,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf/httpd.conf,SubPath:httpd.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf.d/ssl.conf,SubPath:ssl.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:run-httpd,ReadOnly:false,MountPath:/run/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:log-httpd,ReadOnly:false,MountPath:/var/log/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cs95k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(9587d64e-33e3-45d2-8ee7-a776fcc60d88): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 01 09:52:56 crc kubenswrapper[4787]: E1001 09:52:56.084481 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"proxy-httpd\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"]" pod="openstack/ceilometer-0" podUID="9587d64e-33e3-45d2-8ee7-a776fcc60d88" Oct 01 09:52:56 crc kubenswrapper[4787]: I1001 09:52:56.536811 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a63d659d-52a4-4f33-85c2-1461423776a3" path="/var/lib/kubelet/pods/a63d659d-52a4-4f33-85c2-1461423776a3/volumes" Oct 01 09:52:56 crc kubenswrapper[4787]: I1001 09:52:56.691008 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-67879968b-d5598" Oct 01 09:52:56 crc kubenswrapper[4787]: I1001 09:52:56.697338 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7bb97fccb5-vblx7" podUID="a63d659d-52a4-4f33-85c2-1461423776a3" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.138:5353: i/o timeout" Oct 01 09:52:56 crc kubenswrapper[4787]: I1001 09:52:56.867224 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-c4db58965-h4fwp"] Oct 01 09:52:56 crc kubenswrapper[4787]: E1001 09:52:56.868765 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18f38986-fd49-4f82-ad99-0a73264877a6" containerName="neutron-db-sync" Oct 01 09:52:56 crc kubenswrapper[4787]: I1001 09:52:56.868789 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="18f38986-fd49-4f82-ad99-0a73264877a6" containerName="neutron-db-sync" Oct 01 09:52:56 crc kubenswrapper[4787]: E1001 09:52:56.868820 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a63d659d-52a4-4f33-85c2-1461423776a3" containerName="dnsmasq-dns" Oct 01 09:52:56 crc kubenswrapper[4787]: I1001 09:52:56.868827 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="a63d659d-52a4-4f33-85c2-1461423776a3" containerName="dnsmasq-dns" Oct 01 09:52:56 crc kubenswrapper[4787]: E1001 09:52:56.868837 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a63d659d-52a4-4f33-85c2-1461423776a3" containerName="init" Oct 01 09:52:56 crc kubenswrapper[4787]: I1001 09:52:56.868844 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="a63d659d-52a4-4f33-85c2-1461423776a3" containerName="init" Oct 01 09:52:56 crc kubenswrapper[4787]: I1001 09:52:56.869009 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="18f38986-fd49-4f82-ad99-0a73264877a6" containerName="neutron-db-sync" Oct 01 09:52:56 crc kubenswrapper[4787]: I1001 09:52:56.869029 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="a63d659d-52a4-4f33-85c2-1461423776a3" containerName="dnsmasq-dns" Oct 01 09:52:56 crc kubenswrapper[4787]: I1001 09:52:56.870877 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c4db58965-h4fwp" Oct 01 09:52:56 crc kubenswrapper[4787]: I1001 09:52:56.906977 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-c4db58965-h4fwp"] Oct 01 09:52:56 crc kubenswrapper[4787]: I1001 09:52:56.920314 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a7835443-3da0-4c3f-8643-b07cbf790a17-dns-swift-storage-0\") pod \"dnsmasq-dns-c4db58965-h4fwp\" (UID: \"a7835443-3da0-4c3f-8643-b07cbf790a17\") " pod="openstack/dnsmasq-dns-c4db58965-h4fwp" Oct 01 09:52:56 crc kubenswrapper[4787]: I1001 09:52:56.920376 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7835443-3da0-4c3f-8643-b07cbf790a17-config\") pod \"dnsmasq-dns-c4db58965-h4fwp\" (UID: \"a7835443-3da0-4c3f-8643-b07cbf790a17\") " pod="openstack/dnsmasq-dns-c4db58965-h4fwp" Oct 01 09:52:56 crc kubenswrapper[4787]: I1001 09:52:56.920451 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49kwv\" (UniqueName: \"kubernetes.io/projected/a7835443-3da0-4c3f-8643-b07cbf790a17-kube-api-access-49kwv\") pod \"dnsmasq-dns-c4db58965-h4fwp\" (UID: \"a7835443-3da0-4c3f-8643-b07cbf790a17\") " pod="openstack/dnsmasq-dns-c4db58965-h4fwp" Oct 01 09:52:56 crc kubenswrapper[4787]: I1001 09:52:56.920484 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7835443-3da0-4c3f-8643-b07cbf790a17-ovsdbserver-nb\") pod \"dnsmasq-dns-c4db58965-h4fwp\" (UID: \"a7835443-3da0-4c3f-8643-b07cbf790a17\") " pod="openstack/dnsmasq-dns-c4db58965-h4fwp" Oct 01 09:52:56 crc kubenswrapper[4787]: I1001 09:52:56.920526 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7835443-3da0-4c3f-8643-b07cbf790a17-ovsdbserver-sb\") pod \"dnsmasq-dns-c4db58965-h4fwp\" (UID: \"a7835443-3da0-4c3f-8643-b07cbf790a17\") " pod="openstack/dnsmasq-dns-c4db58965-h4fwp" Oct 01 09:52:56 crc kubenswrapper[4787]: I1001 09:52:56.920549 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7835443-3da0-4c3f-8643-b07cbf790a17-dns-svc\") pod \"dnsmasq-dns-c4db58965-h4fwp\" (UID: \"a7835443-3da0-4c3f-8643-b07cbf790a17\") " pod="openstack/dnsmasq-dns-c4db58965-h4fwp" Oct 01 09:52:56 crc kubenswrapper[4787]: I1001 09:52:56.982289 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-858fdb94fd-6tpq6"] Oct 01 09:52:56 crc kubenswrapper[4787]: I1001 09:52:56.983902 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-858fdb94fd-6tpq6" Oct 01 09:52:56 crc kubenswrapper[4787]: I1001 09:52:56.987184 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 01 09:52:56 crc kubenswrapper[4787]: I1001 09:52:56.987386 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 01 09:52:56 crc kubenswrapper[4787]: I1001 09:52:56.987563 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-jgvxn" Oct 01 09:52:56 crc kubenswrapper[4787]: I1001 09:52:56.987834 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 01 09:52:56 crc kubenswrapper[4787]: I1001 09:52:56.995120 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-858fdb94fd-6tpq6"] Oct 01 09:52:57 crc kubenswrapper[4787]: I1001 09:52:57.022042 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0ac97b66-7233-4fcc-9b0f-856a67a62b4d-config\") pod \"neutron-858fdb94fd-6tpq6\" (UID: \"0ac97b66-7233-4fcc-9b0f-856a67a62b4d\") " pod="openstack/neutron-858fdb94fd-6tpq6" Oct 01 09:52:57 crc kubenswrapper[4787]: I1001 09:52:57.022155 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0ac97b66-7233-4fcc-9b0f-856a67a62b4d-httpd-config\") pod \"neutron-858fdb94fd-6tpq6\" (UID: \"0ac97b66-7233-4fcc-9b0f-856a67a62b4d\") " pod="openstack/neutron-858fdb94fd-6tpq6" Oct 01 09:52:57 crc kubenswrapper[4787]: I1001 09:52:57.022202 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a7835443-3da0-4c3f-8643-b07cbf790a17-dns-swift-storage-0\") pod \"dnsmasq-dns-c4db58965-h4fwp\" (UID: \"a7835443-3da0-4c3f-8643-b07cbf790a17\") " pod="openstack/dnsmasq-dns-c4db58965-h4fwp" Oct 01 09:52:57 crc kubenswrapper[4787]: I1001 09:52:57.022258 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7835443-3da0-4c3f-8643-b07cbf790a17-config\") pod \"dnsmasq-dns-c4db58965-h4fwp\" (UID: \"a7835443-3da0-4c3f-8643-b07cbf790a17\") " pod="openstack/dnsmasq-dns-c4db58965-h4fwp" Oct 01 09:52:57 crc kubenswrapper[4787]: I1001 09:52:57.022285 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lq7kt\" (UniqueName: \"kubernetes.io/projected/0ac97b66-7233-4fcc-9b0f-856a67a62b4d-kube-api-access-lq7kt\") pod \"neutron-858fdb94fd-6tpq6\" (UID: \"0ac97b66-7233-4fcc-9b0f-856a67a62b4d\") " pod="openstack/neutron-858fdb94fd-6tpq6" Oct 01 09:52:57 crc kubenswrapper[4787]: I1001 09:52:57.023253 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7835443-3da0-4c3f-8643-b07cbf790a17-config\") pod \"dnsmasq-dns-c4db58965-h4fwp\" (UID: \"a7835443-3da0-4c3f-8643-b07cbf790a17\") " pod="openstack/dnsmasq-dns-c4db58965-h4fwp" Oct 01 09:52:57 crc kubenswrapper[4787]: I1001 09:52:57.023258 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a7835443-3da0-4c3f-8643-b07cbf790a17-dns-swift-storage-0\") pod \"dnsmasq-dns-c4db58965-h4fwp\" (UID: \"a7835443-3da0-4c3f-8643-b07cbf790a17\") " pod="openstack/dnsmasq-dns-c4db58965-h4fwp" Oct 01 09:52:57 crc kubenswrapper[4787]: I1001 09:52:57.023326 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ac97b66-7233-4fcc-9b0f-856a67a62b4d-combined-ca-bundle\") pod \"neutron-858fdb94fd-6tpq6\" (UID: \"0ac97b66-7233-4fcc-9b0f-856a67a62b4d\") " pod="openstack/neutron-858fdb94fd-6tpq6" Oct 01 09:52:57 crc kubenswrapper[4787]: I1001 09:52:57.023507 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49kwv\" (UniqueName: \"kubernetes.io/projected/a7835443-3da0-4c3f-8643-b07cbf790a17-kube-api-access-49kwv\") pod \"dnsmasq-dns-c4db58965-h4fwp\" (UID: \"a7835443-3da0-4c3f-8643-b07cbf790a17\") " pod="openstack/dnsmasq-dns-c4db58965-h4fwp" Oct 01 09:52:57 crc kubenswrapper[4787]: I1001 09:52:57.023574 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7835443-3da0-4c3f-8643-b07cbf790a17-ovsdbserver-nb\") pod \"dnsmasq-dns-c4db58965-h4fwp\" (UID: \"a7835443-3da0-4c3f-8643-b07cbf790a17\") " pod="openstack/dnsmasq-dns-c4db58965-h4fwp" Oct 01 09:52:57 crc kubenswrapper[4787]: I1001 09:52:57.023629 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ac97b66-7233-4fcc-9b0f-856a67a62b4d-ovndb-tls-certs\") pod \"neutron-858fdb94fd-6tpq6\" (UID: \"0ac97b66-7233-4fcc-9b0f-856a67a62b4d\") " pod="openstack/neutron-858fdb94fd-6tpq6" Oct 01 09:52:57 crc kubenswrapper[4787]: I1001 09:52:57.023685 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7835443-3da0-4c3f-8643-b07cbf790a17-ovsdbserver-sb\") pod \"dnsmasq-dns-c4db58965-h4fwp\" (UID: \"a7835443-3da0-4c3f-8643-b07cbf790a17\") " pod="openstack/dnsmasq-dns-c4db58965-h4fwp" Oct 01 09:52:57 crc kubenswrapper[4787]: I1001 09:52:57.023724 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7835443-3da0-4c3f-8643-b07cbf790a17-dns-svc\") pod \"dnsmasq-dns-c4db58965-h4fwp\" (UID: \"a7835443-3da0-4c3f-8643-b07cbf790a17\") " pod="openstack/dnsmasq-dns-c4db58965-h4fwp" Oct 01 09:52:57 crc kubenswrapper[4787]: I1001 09:52:57.024347 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7835443-3da0-4c3f-8643-b07cbf790a17-dns-svc\") pod \"dnsmasq-dns-c4db58965-h4fwp\" (UID: \"a7835443-3da0-4c3f-8643-b07cbf790a17\") " pod="openstack/dnsmasq-dns-c4db58965-h4fwp" Oct 01 09:52:57 crc kubenswrapper[4787]: I1001 09:52:57.025194 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7835443-3da0-4c3f-8643-b07cbf790a17-ovsdbserver-sb\") pod \"dnsmasq-dns-c4db58965-h4fwp\" (UID: \"a7835443-3da0-4c3f-8643-b07cbf790a17\") " pod="openstack/dnsmasq-dns-c4db58965-h4fwp" Oct 01 09:52:57 crc kubenswrapper[4787]: I1001 09:52:57.026236 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7835443-3da0-4c3f-8643-b07cbf790a17-ovsdbserver-nb\") pod \"dnsmasq-dns-c4db58965-h4fwp\" (UID: \"a7835443-3da0-4c3f-8643-b07cbf790a17\") " pod="openstack/dnsmasq-dns-c4db58965-h4fwp" Oct 01 09:52:57 crc kubenswrapper[4787]: I1001 09:52:57.046840 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49kwv\" (UniqueName: \"kubernetes.io/projected/a7835443-3da0-4c3f-8643-b07cbf790a17-kube-api-access-49kwv\") pod \"dnsmasq-dns-c4db58965-h4fwp\" (UID: \"a7835443-3da0-4c3f-8643-b07cbf790a17\") " pod="openstack/dnsmasq-dns-c4db58965-h4fwp" Oct 01 09:52:57 crc kubenswrapper[4787]: I1001 09:52:57.068415 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7d65d5957b-44rtv" Oct 01 09:52:57 crc kubenswrapper[4787]: I1001 09:52:57.070703 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-wndd6" event={"ID":"15111beb-b05b-440f-a4c9-077eca1c37d1","Type":"ContainerStarted","Data":"77f44314acff33e7ec38659c6b9db47146a92cd1eb9a9366d6be65d8a7e8d8ce"} Oct 01 09:52:57 crc kubenswrapper[4787]: I1001 09:52:57.070950 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9587d64e-33e3-45d2-8ee7-a776fcc60d88" containerName="sg-core" containerID="cri-o://8b380d576833d002f4a4c98eff2e55905be1be5916394c71d0684713baac85ef" gracePeriod=30 Oct 01 09:52:57 crc kubenswrapper[4787]: I1001 09:52:57.070936 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9587d64e-33e3-45d2-8ee7-a776fcc60d88" containerName="ceilometer-notification-agent" containerID="cri-o://7564cf5c7eb4e3180d068e49673d5fb7e6cd1d11878d243edfaee1b52031f33d" gracePeriod=30 Oct 01 09:52:57 crc kubenswrapper[4787]: I1001 09:52:57.121409 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-wndd6" podStartSLOduration=3.173703447 podStartE2EDuration="29.121386349s" podCreationTimestamp="2025-10-01 09:52:28 +0000 UTC" firstStartedPulling="2025-10-01 09:52:30.132656117 +0000 UTC m=+982.247800274" lastFinishedPulling="2025-10-01 09:52:56.080339029 +0000 UTC m=+1008.195483176" observedRunningTime="2025-10-01 09:52:57.114941506 +0000 UTC m=+1009.230085663" watchObservedRunningTime="2025-10-01 09:52:57.121386349 +0000 UTC m=+1009.236530516" Oct 01 09:52:57 crc kubenswrapper[4787]: I1001 09:52:57.126769 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ac97b66-7233-4fcc-9b0f-856a67a62b4d-ovndb-tls-certs\") pod \"neutron-858fdb94fd-6tpq6\" (UID: \"0ac97b66-7233-4fcc-9b0f-856a67a62b4d\") " pod="openstack/neutron-858fdb94fd-6tpq6" Oct 01 09:52:57 crc kubenswrapper[4787]: I1001 09:52:57.126861 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0ac97b66-7233-4fcc-9b0f-856a67a62b4d-config\") pod \"neutron-858fdb94fd-6tpq6\" (UID: \"0ac97b66-7233-4fcc-9b0f-856a67a62b4d\") " pod="openstack/neutron-858fdb94fd-6tpq6" Oct 01 09:52:57 crc kubenswrapper[4787]: I1001 09:52:57.126887 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0ac97b66-7233-4fcc-9b0f-856a67a62b4d-httpd-config\") pod \"neutron-858fdb94fd-6tpq6\" (UID: \"0ac97b66-7233-4fcc-9b0f-856a67a62b4d\") " pod="openstack/neutron-858fdb94fd-6tpq6" Oct 01 09:52:57 crc kubenswrapper[4787]: I1001 09:52:57.127003 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lq7kt\" (UniqueName: \"kubernetes.io/projected/0ac97b66-7233-4fcc-9b0f-856a67a62b4d-kube-api-access-lq7kt\") pod \"neutron-858fdb94fd-6tpq6\" (UID: \"0ac97b66-7233-4fcc-9b0f-856a67a62b4d\") " pod="openstack/neutron-858fdb94fd-6tpq6" Oct 01 09:52:57 crc kubenswrapper[4787]: I1001 09:52:57.127027 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ac97b66-7233-4fcc-9b0f-856a67a62b4d-combined-ca-bundle\") pod \"neutron-858fdb94fd-6tpq6\" (UID: \"0ac97b66-7233-4fcc-9b0f-856a67a62b4d\") " pod="openstack/neutron-858fdb94fd-6tpq6" Oct 01 09:52:57 crc kubenswrapper[4787]: I1001 09:52:57.138465 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ac97b66-7233-4fcc-9b0f-856a67a62b4d-combined-ca-bundle\") pod \"neutron-858fdb94fd-6tpq6\" (UID: \"0ac97b66-7233-4fcc-9b0f-856a67a62b4d\") " pod="openstack/neutron-858fdb94fd-6tpq6" Oct 01 09:52:57 crc kubenswrapper[4787]: I1001 09:52:57.141940 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ac97b66-7233-4fcc-9b0f-856a67a62b4d-ovndb-tls-certs\") pod \"neutron-858fdb94fd-6tpq6\" (UID: \"0ac97b66-7233-4fcc-9b0f-856a67a62b4d\") " pod="openstack/neutron-858fdb94fd-6tpq6" Oct 01 09:52:57 crc kubenswrapper[4787]: I1001 09:52:57.147257 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/0ac97b66-7233-4fcc-9b0f-856a67a62b4d-config\") pod \"neutron-858fdb94fd-6tpq6\" (UID: \"0ac97b66-7233-4fcc-9b0f-856a67a62b4d\") " pod="openstack/neutron-858fdb94fd-6tpq6" Oct 01 09:52:57 crc kubenswrapper[4787]: I1001 09:52:57.151066 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-67879968b-d5598"] Oct 01 09:52:57 crc kubenswrapper[4787]: I1001 09:52:57.151398 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-67879968b-d5598" podUID="7983e213-c317-42b2-8944-85bdb78dac8f" containerName="horizon" containerID="cri-o://c363923f03c8431731ce61fb7088f6a51538ab7d45b2f5fb59d3d2d245ae02f0" gracePeriod=30 Oct 01 09:52:57 crc kubenswrapper[4787]: I1001 09:52:57.151343 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-67879968b-d5598" podUID="7983e213-c317-42b2-8944-85bdb78dac8f" containerName="horizon-log" containerID="cri-o://37b1a9eaf9bc59c2abf28488f9ac98956d1e78ecc9bdf3f5d70a1af066d4e549" gracePeriod=30 Oct 01 09:52:57 crc kubenswrapper[4787]: I1001 09:52:57.157772 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0ac97b66-7233-4fcc-9b0f-856a67a62b4d-httpd-config\") pod \"neutron-858fdb94fd-6tpq6\" (UID: \"0ac97b66-7233-4fcc-9b0f-856a67a62b4d\") " pod="openstack/neutron-858fdb94fd-6tpq6" Oct 01 09:52:57 crc kubenswrapper[4787]: I1001 09:52:57.166108 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lq7kt\" (UniqueName: \"kubernetes.io/projected/0ac97b66-7233-4fcc-9b0f-856a67a62b4d-kube-api-access-lq7kt\") pod \"neutron-858fdb94fd-6tpq6\" (UID: \"0ac97b66-7233-4fcc-9b0f-856a67a62b4d\") " pod="openstack/neutron-858fdb94fd-6tpq6" Oct 01 09:52:57 crc kubenswrapper[4787]: I1001 09:52:57.212583 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c4db58965-h4fwp" Oct 01 09:52:57 crc kubenswrapper[4787]: I1001 09:52:57.302112 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-858fdb94fd-6tpq6" Oct 01 09:52:57 crc kubenswrapper[4787]: I1001 09:52:57.735793 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-c4db58965-h4fwp"] Oct 01 09:52:57 crc kubenswrapper[4787]: W1001 09:52:57.739942 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda7835443_3da0_4c3f_8643_b07cbf790a17.slice/crio-c1bc7b380d40fc56ea801fca38d39015451179ce2f681cd59fdd6014a7d8f74e WatchSource:0}: Error finding container c1bc7b380d40fc56ea801fca38d39015451179ce2f681cd59fdd6014a7d8f74e: Status 404 returned error can't find the container with id c1bc7b380d40fc56ea801fca38d39015451179ce2f681cd59fdd6014a7d8f74e Oct 01 09:52:58 crc kubenswrapper[4787]: I1001 09:52:58.081190 4787 generic.go:334] "Generic (PLEG): container finished" podID="a7835443-3da0-4c3f-8643-b07cbf790a17" containerID="cfb994f04e37d89a8d2e9681493c188d01b7b7dad216fc2587b4419c5a95e425" exitCode=0 Oct 01 09:52:58 crc kubenswrapper[4787]: I1001 09:52:58.081324 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c4db58965-h4fwp" event={"ID":"a7835443-3da0-4c3f-8643-b07cbf790a17","Type":"ContainerDied","Data":"cfb994f04e37d89a8d2e9681493c188d01b7b7dad216fc2587b4419c5a95e425"} Oct 01 09:52:58 crc kubenswrapper[4787]: I1001 09:52:58.081579 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c4db58965-h4fwp" event={"ID":"a7835443-3da0-4c3f-8643-b07cbf790a17","Type":"ContainerStarted","Data":"c1bc7b380d40fc56ea801fca38d39015451179ce2f681cd59fdd6014a7d8f74e"} Oct 01 09:52:58 crc kubenswrapper[4787]: I1001 09:52:58.085575 4787 generic.go:334] "Generic (PLEG): container finished" podID="9587d64e-33e3-45d2-8ee7-a776fcc60d88" containerID="8b380d576833d002f4a4c98eff2e55905be1be5916394c71d0684713baac85ef" exitCode=2 Oct 01 09:52:58 crc kubenswrapper[4787]: I1001 09:52:58.085647 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9587d64e-33e3-45d2-8ee7-a776fcc60d88","Type":"ContainerDied","Data":"8b380d576833d002f4a4c98eff2e55905be1be5916394c71d0684713baac85ef"} Oct 01 09:52:58 crc kubenswrapper[4787]: I1001 09:52:58.250203 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-858fdb94fd-6tpq6"] Oct 01 09:52:58 crc kubenswrapper[4787]: I1001 09:52:58.898665 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:52:58 crc kubenswrapper[4787]: I1001 09:52:58.961879 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9587d64e-33e3-45d2-8ee7-a776fcc60d88-sg-core-conf-yaml\") pod \"9587d64e-33e3-45d2-8ee7-a776fcc60d88\" (UID: \"9587d64e-33e3-45d2-8ee7-a776fcc60d88\") " Oct 01 09:52:58 crc kubenswrapper[4787]: I1001 09:52:58.961932 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9587d64e-33e3-45d2-8ee7-a776fcc60d88-run-httpd\") pod \"9587d64e-33e3-45d2-8ee7-a776fcc60d88\" (UID: \"9587d64e-33e3-45d2-8ee7-a776fcc60d88\") " Oct 01 09:52:58 crc kubenswrapper[4787]: I1001 09:52:58.961971 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9587d64e-33e3-45d2-8ee7-a776fcc60d88-config-data\") pod \"9587d64e-33e3-45d2-8ee7-a776fcc60d88\" (UID: \"9587d64e-33e3-45d2-8ee7-a776fcc60d88\") " Oct 01 09:52:58 crc kubenswrapper[4787]: I1001 09:52:58.962049 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9587d64e-33e3-45d2-8ee7-a776fcc60d88-combined-ca-bundle\") pod \"9587d64e-33e3-45d2-8ee7-a776fcc60d88\" (UID: \"9587d64e-33e3-45d2-8ee7-a776fcc60d88\") " Oct 01 09:52:58 crc kubenswrapper[4787]: I1001 09:52:58.962066 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9587d64e-33e3-45d2-8ee7-a776fcc60d88-log-httpd\") pod \"9587d64e-33e3-45d2-8ee7-a776fcc60d88\" (UID: \"9587d64e-33e3-45d2-8ee7-a776fcc60d88\") " Oct 01 09:52:58 crc kubenswrapper[4787]: I1001 09:52:58.962145 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9587d64e-33e3-45d2-8ee7-a776fcc60d88-scripts\") pod \"9587d64e-33e3-45d2-8ee7-a776fcc60d88\" (UID: \"9587d64e-33e3-45d2-8ee7-a776fcc60d88\") " Oct 01 09:52:58 crc kubenswrapper[4787]: I1001 09:52:58.962203 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cs95k\" (UniqueName: \"kubernetes.io/projected/9587d64e-33e3-45d2-8ee7-a776fcc60d88-kube-api-access-cs95k\") pod \"9587d64e-33e3-45d2-8ee7-a776fcc60d88\" (UID: \"9587d64e-33e3-45d2-8ee7-a776fcc60d88\") " Oct 01 09:52:58 crc kubenswrapper[4787]: I1001 09:52:58.962388 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9587d64e-33e3-45d2-8ee7-a776fcc60d88-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9587d64e-33e3-45d2-8ee7-a776fcc60d88" (UID: "9587d64e-33e3-45d2-8ee7-a776fcc60d88"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:52:58 crc kubenswrapper[4787]: I1001 09:52:58.962553 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9587d64e-33e3-45d2-8ee7-a776fcc60d88-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9587d64e-33e3-45d2-8ee7-a776fcc60d88" (UID: "9587d64e-33e3-45d2-8ee7-a776fcc60d88"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:52:58 crc kubenswrapper[4787]: I1001 09:52:58.962767 4787 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9587d64e-33e3-45d2-8ee7-a776fcc60d88-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:58 crc kubenswrapper[4787]: I1001 09:52:58.962785 4787 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9587d64e-33e3-45d2-8ee7-a776fcc60d88-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:58 crc kubenswrapper[4787]: I1001 09:52:58.972233 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9587d64e-33e3-45d2-8ee7-a776fcc60d88-kube-api-access-cs95k" (OuterVolumeSpecName: "kube-api-access-cs95k") pod "9587d64e-33e3-45d2-8ee7-a776fcc60d88" (UID: "9587d64e-33e3-45d2-8ee7-a776fcc60d88"). InnerVolumeSpecName "kube-api-access-cs95k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:52:58 crc kubenswrapper[4787]: I1001 09:52:58.979797 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9587d64e-33e3-45d2-8ee7-a776fcc60d88-scripts" (OuterVolumeSpecName: "scripts") pod "9587d64e-33e3-45d2-8ee7-a776fcc60d88" (UID: "9587d64e-33e3-45d2-8ee7-a776fcc60d88"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:52:58 crc kubenswrapper[4787]: I1001 09:52:58.990895 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-84d5f4684d-p2xjn" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.004334 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9587d64e-33e3-45d2-8ee7-a776fcc60d88-config-data" (OuterVolumeSpecName: "config-data") pod "9587d64e-33e3-45d2-8ee7-a776fcc60d88" (UID: "9587d64e-33e3-45d2-8ee7-a776fcc60d88"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.007166 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9587d64e-33e3-45d2-8ee7-a776fcc60d88-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "9587d64e-33e3-45d2-8ee7-a776fcc60d88" (UID: "9587d64e-33e3-45d2-8ee7-a776fcc60d88"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.010217 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9587d64e-33e3-45d2-8ee7-a776fcc60d88-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9587d64e-33e3-45d2-8ee7-a776fcc60d88" (UID: "9587d64e-33e3-45d2-8ee7-a776fcc60d88"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.064936 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9587d64e-33e3-45d2-8ee7-a776fcc60d88-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.064971 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cs95k\" (UniqueName: \"kubernetes.io/projected/9587d64e-33e3-45d2-8ee7-a776fcc60d88-kube-api-access-cs95k\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.064984 4787 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9587d64e-33e3-45d2-8ee7-a776fcc60d88-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.064993 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9587d64e-33e3-45d2-8ee7-a776fcc60d88-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.065016 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9587d64e-33e3-45d2-8ee7-a776fcc60d88-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.103824 4787 generic.go:334] "Generic (PLEG): container finished" podID="9587d64e-33e3-45d2-8ee7-a776fcc60d88" containerID="7564cf5c7eb4e3180d068e49673d5fb7e6cd1d11878d243edfaee1b52031f33d" exitCode=0 Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.103952 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9587d64e-33e3-45d2-8ee7-a776fcc60d88","Type":"ContainerDied","Data":"7564cf5c7eb4e3180d068e49673d5fb7e6cd1d11878d243edfaee1b52031f33d"} Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.103989 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9587d64e-33e3-45d2-8ee7-a776fcc60d88","Type":"ContainerDied","Data":"3a332e3f42bdb96a3a7da96a0f57511b9d0dde1414884f7ff47a9d91125a8639"} Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.104022 4787 scope.go:117] "RemoveContainer" containerID="8b380d576833d002f4a4c98eff2e55905be1be5916394c71d0684713baac85ef" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.109226 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.116141 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c4db58965-h4fwp" event={"ID":"a7835443-3da0-4c3f-8643-b07cbf790a17","Type":"ContainerStarted","Data":"f604cceb35751cdf4ced9c10d851eef06a4de09e5f2c818aa1f04a04d70ef53c"} Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.117578 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-c4db58965-h4fwp" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.118035 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-858fdb94fd-6tpq6" event={"ID":"0ac97b66-7233-4fcc-9b0f-856a67a62b4d","Type":"ContainerStarted","Data":"2b5867e80774c39c9e0b986b908a0d57bfbab946cc4e15e39311fc5003ff06a0"} Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.118056 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-858fdb94fd-6tpq6" event={"ID":"0ac97b66-7233-4fcc-9b0f-856a67a62b4d","Type":"ContainerStarted","Data":"09101ceda6bf2ba07365f993acdcbbb18dd498f848daa46e32c04bb05065a77b"} Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.118069 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-858fdb94fd-6tpq6" event={"ID":"0ac97b66-7233-4fcc-9b0f-856a67a62b4d","Type":"ContainerStarted","Data":"879a9663415d32abcb590d620ef91ad608f4b823f66c8ca7a06051798ee1e67d"} Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.118673 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-858fdb94fd-6tpq6" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.170451 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-858fdb94fd-6tpq6" podStartSLOduration=3.170428303 podStartE2EDuration="3.170428303s" podCreationTimestamp="2025-10-01 09:52:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:52:59.144293151 +0000 UTC m=+1011.259437308" watchObservedRunningTime="2025-10-01 09:52:59.170428303 +0000 UTC m=+1011.285572460" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.205395 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-c4db58965-h4fwp" podStartSLOduration=3.205370658 podStartE2EDuration="3.205370658s" podCreationTimestamp="2025-10-01 09:52:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:52:59.187825313 +0000 UTC m=+1011.302969480" watchObservedRunningTime="2025-10-01 09:52:59.205370658 +0000 UTC m=+1011.320514805" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.247365 4787 scope.go:117] "RemoveContainer" containerID="7564cf5c7eb4e3180d068e49673d5fb7e6cd1d11878d243edfaee1b52031f33d" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.314269 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.321228 4787 scope.go:117] "RemoveContainer" containerID="8b380d576833d002f4a4c98eff2e55905be1be5916394c71d0684713baac85ef" Oct 01 09:52:59 crc kubenswrapper[4787]: E1001 09:52:59.321658 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b380d576833d002f4a4c98eff2e55905be1be5916394c71d0684713baac85ef\": container with ID starting with 8b380d576833d002f4a4c98eff2e55905be1be5916394c71d0684713baac85ef not found: ID does not exist" containerID="8b380d576833d002f4a4c98eff2e55905be1be5916394c71d0684713baac85ef" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.321686 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b380d576833d002f4a4c98eff2e55905be1be5916394c71d0684713baac85ef"} err="failed to get container status \"8b380d576833d002f4a4c98eff2e55905be1be5916394c71d0684713baac85ef\": rpc error: code = NotFound desc = could not find container \"8b380d576833d002f4a4c98eff2e55905be1be5916394c71d0684713baac85ef\": container with ID starting with 8b380d576833d002f4a4c98eff2e55905be1be5916394c71d0684713baac85ef not found: ID does not exist" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.321707 4787 scope.go:117] "RemoveContainer" containerID="7564cf5c7eb4e3180d068e49673d5fb7e6cd1d11878d243edfaee1b52031f33d" Oct 01 09:52:59 crc kubenswrapper[4787]: E1001 09:52:59.321869 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7564cf5c7eb4e3180d068e49673d5fb7e6cd1d11878d243edfaee1b52031f33d\": container with ID starting with 7564cf5c7eb4e3180d068e49673d5fb7e6cd1d11878d243edfaee1b52031f33d not found: ID does not exist" containerID="7564cf5c7eb4e3180d068e49673d5fb7e6cd1d11878d243edfaee1b52031f33d" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.321883 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7564cf5c7eb4e3180d068e49673d5fb7e6cd1d11878d243edfaee1b52031f33d"} err="failed to get container status \"7564cf5c7eb4e3180d068e49673d5fb7e6cd1d11878d243edfaee1b52031f33d\": rpc error: code = NotFound desc = could not find container \"7564cf5c7eb4e3180d068e49673d5fb7e6cd1d11878d243edfaee1b52031f33d\": container with ID starting with 7564cf5c7eb4e3180d068e49673d5fb7e6cd1d11878d243edfaee1b52031f33d not found: ID does not exist" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.326295 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.338826 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:52:59 crc kubenswrapper[4787]: E1001 09:52:59.339407 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9587d64e-33e3-45d2-8ee7-a776fcc60d88" containerName="ceilometer-notification-agent" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.339575 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="9587d64e-33e3-45d2-8ee7-a776fcc60d88" containerName="ceilometer-notification-agent" Oct 01 09:52:59 crc kubenswrapper[4787]: E1001 09:52:59.339814 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9587d64e-33e3-45d2-8ee7-a776fcc60d88" containerName="sg-core" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.339978 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="9587d64e-33e3-45d2-8ee7-a776fcc60d88" containerName="sg-core" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.340331 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="9587d64e-33e3-45d2-8ee7-a776fcc60d88" containerName="sg-core" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.340883 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="9587d64e-33e3-45d2-8ee7-a776fcc60d88" containerName="ceilometer-notification-agent" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.344773 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.347166 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.347490 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.359874 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.414502 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7zvd\" (UniqueName: \"kubernetes.io/projected/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-kube-api-access-j7zvd\") pod \"ceilometer-0\" (UID: \"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c\") " pod="openstack/ceilometer-0" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.414560 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c\") " pod="openstack/ceilometer-0" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.414620 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-scripts\") pod \"ceilometer-0\" (UID: \"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c\") " pod="openstack/ceilometer-0" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.414851 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c\") " pod="openstack/ceilometer-0" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.414878 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-config-data\") pod \"ceilometer-0\" (UID: \"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c\") " pod="openstack/ceilometer-0" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.414921 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-log-httpd\") pod \"ceilometer-0\" (UID: \"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c\") " pod="openstack/ceilometer-0" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.414940 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-run-httpd\") pod \"ceilometer-0\" (UID: \"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c\") " pod="openstack/ceilometer-0" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.518209 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-scripts\") pod \"ceilometer-0\" (UID: \"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c\") " pod="openstack/ceilometer-0" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.518309 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c\") " pod="openstack/ceilometer-0" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.518334 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-config-data\") pod \"ceilometer-0\" (UID: \"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c\") " pod="openstack/ceilometer-0" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.518375 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-log-httpd\") pod \"ceilometer-0\" (UID: \"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c\") " pod="openstack/ceilometer-0" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.518393 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-run-httpd\") pod \"ceilometer-0\" (UID: \"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c\") " pod="openstack/ceilometer-0" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.518429 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7zvd\" (UniqueName: \"kubernetes.io/projected/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-kube-api-access-j7zvd\") pod \"ceilometer-0\" (UID: \"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c\") " pod="openstack/ceilometer-0" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.518462 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c\") " pod="openstack/ceilometer-0" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.523502 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-run-httpd\") pod \"ceilometer-0\" (UID: \"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c\") " pod="openstack/ceilometer-0" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.523862 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-log-httpd\") pod \"ceilometer-0\" (UID: \"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c\") " pod="openstack/ceilometer-0" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.529579 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-scripts\") pod \"ceilometer-0\" (UID: \"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c\") " pod="openstack/ceilometer-0" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.529785 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c\") " pod="openstack/ceilometer-0" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.548901 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-config-data\") pod \"ceilometer-0\" (UID: \"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c\") " pod="openstack/ceilometer-0" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.550540 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c\") " pod="openstack/ceilometer-0" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.554663 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7zvd\" (UniqueName: \"kubernetes.io/projected/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-kube-api-access-j7zvd\") pod \"ceilometer-0\" (UID: \"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c\") " pod="openstack/ceilometer-0" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.679975 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.923897 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-58d9cd74f7-86sts"] Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.925742 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-58d9cd74f7-86sts" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.927752 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.927942 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 01 09:52:59 crc kubenswrapper[4787]: I1001 09:52:59.983400 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-58d9cd74f7-86sts"] Oct 01 09:53:00 crc kubenswrapper[4787]: I1001 09:53:00.026474 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0-config\") pod \"neutron-58d9cd74f7-86sts\" (UID: \"1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0\") " pod="openstack/neutron-58d9cd74f7-86sts" Oct 01 09:53:00 crc kubenswrapper[4787]: I1001 09:53:00.026529 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0-public-tls-certs\") pod \"neutron-58d9cd74f7-86sts\" (UID: \"1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0\") " pod="openstack/neutron-58d9cd74f7-86sts" Oct 01 09:53:00 crc kubenswrapper[4787]: I1001 09:53:00.026551 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0-ovndb-tls-certs\") pod \"neutron-58d9cd74f7-86sts\" (UID: \"1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0\") " pod="openstack/neutron-58d9cd74f7-86sts" Oct 01 09:53:00 crc kubenswrapper[4787]: I1001 09:53:00.026749 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0-httpd-config\") pod \"neutron-58d9cd74f7-86sts\" (UID: \"1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0\") " pod="openstack/neutron-58d9cd74f7-86sts" Oct 01 09:53:00 crc kubenswrapper[4787]: I1001 09:53:00.026862 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0-internal-tls-certs\") pod \"neutron-58d9cd74f7-86sts\" (UID: \"1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0\") " pod="openstack/neutron-58d9cd74f7-86sts" Oct 01 09:53:00 crc kubenswrapper[4787]: I1001 09:53:00.026925 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0-combined-ca-bundle\") pod \"neutron-58d9cd74f7-86sts\" (UID: \"1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0\") " pod="openstack/neutron-58d9cd74f7-86sts" Oct 01 09:53:00 crc kubenswrapper[4787]: I1001 09:53:00.027217 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8r6cs\" (UniqueName: \"kubernetes.io/projected/1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0-kube-api-access-8r6cs\") pod \"neutron-58d9cd74f7-86sts\" (UID: \"1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0\") " pod="openstack/neutron-58d9cd74f7-86sts" Oct 01 09:53:00 crc kubenswrapper[4787]: I1001 09:53:00.129911 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8r6cs\" (UniqueName: \"kubernetes.io/projected/1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0-kube-api-access-8r6cs\") pod \"neutron-58d9cd74f7-86sts\" (UID: \"1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0\") " pod="openstack/neutron-58d9cd74f7-86sts" Oct 01 09:53:00 crc kubenswrapper[4787]: I1001 09:53:00.130038 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0-config\") pod \"neutron-58d9cd74f7-86sts\" (UID: \"1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0\") " pod="openstack/neutron-58d9cd74f7-86sts" Oct 01 09:53:00 crc kubenswrapper[4787]: I1001 09:53:00.130105 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0-public-tls-certs\") pod \"neutron-58d9cd74f7-86sts\" (UID: \"1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0\") " pod="openstack/neutron-58d9cd74f7-86sts" Oct 01 09:53:00 crc kubenswrapper[4787]: I1001 09:53:00.130133 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0-ovndb-tls-certs\") pod \"neutron-58d9cd74f7-86sts\" (UID: \"1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0\") " pod="openstack/neutron-58d9cd74f7-86sts" Oct 01 09:53:00 crc kubenswrapper[4787]: I1001 09:53:00.130188 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0-httpd-config\") pod \"neutron-58d9cd74f7-86sts\" (UID: \"1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0\") " pod="openstack/neutron-58d9cd74f7-86sts" Oct 01 09:53:00 crc kubenswrapper[4787]: I1001 09:53:00.130227 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0-internal-tls-certs\") pod \"neutron-58d9cd74f7-86sts\" (UID: \"1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0\") " pod="openstack/neutron-58d9cd74f7-86sts" Oct 01 09:53:00 crc kubenswrapper[4787]: I1001 09:53:00.130254 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0-combined-ca-bundle\") pod \"neutron-58d9cd74f7-86sts\" (UID: \"1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0\") " pod="openstack/neutron-58d9cd74f7-86sts" Oct 01 09:53:00 crc kubenswrapper[4787]: I1001 09:53:00.135700 4787 generic.go:334] "Generic (PLEG): container finished" podID="15111beb-b05b-440f-a4c9-077eca1c37d1" containerID="77f44314acff33e7ec38659c6b9db47146a92cd1eb9a9366d6be65d8a7e8d8ce" exitCode=0 Oct 01 09:53:00 crc kubenswrapper[4787]: I1001 09:53:00.136158 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-wndd6" event={"ID":"15111beb-b05b-440f-a4c9-077eca1c37d1","Type":"ContainerDied","Data":"77f44314acff33e7ec38659c6b9db47146a92cd1eb9a9366d6be65d8a7e8d8ce"} Oct 01 09:53:00 crc kubenswrapper[4787]: I1001 09:53:00.136839 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0-ovndb-tls-certs\") pod \"neutron-58d9cd74f7-86sts\" (UID: \"1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0\") " pod="openstack/neutron-58d9cd74f7-86sts" Oct 01 09:53:00 crc kubenswrapper[4787]: I1001 09:53:00.137700 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0-httpd-config\") pod \"neutron-58d9cd74f7-86sts\" (UID: \"1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0\") " pod="openstack/neutron-58d9cd74f7-86sts" Oct 01 09:53:00 crc kubenswrapper[4787]: I1001 09:53:00.138184 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0-config\") pod \"neutron-58d9cd74f7-86sts\" (UID: \"1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0\") " pod="openstack/neutron-58d9cd74f7-86sts" Oct 01 09:53:00 crc kubenswrapper[4787]: I1001 09:53:00.145696 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0-combined-ca-bundle\") pod \"neutron-58d9cd74f7-86sts\" (UID: \"1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0\") " pod="openstack/neutron-58d9cd74f7-86sts" Oct 01 09:53:00 crc kubenswrapper[4787]: I1001 09:53:00.149304 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0-public-tls-certs\") pod \"neutron-58d9cd74f7-86sts\" (UID: \"1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0\") " pod="openstack/neutron-58d9cd74f7-86sts" Oct 01 09:53:00 crc kubenswrapper[4787]: I1001 09:53:00.150617 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8r6cs\" (UniqueName: \"kubernetes.io/projected/1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0-kube-api-access-8r6cs\") pod \"neutron-58d9cd74f7-86sts\" (UID: \"1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0\") " pod="openstack/neutron-58d9cd74f7-86sts" Oct 01 09:53:00 crc kubenswrapper[4787]: I1001 09:53:00.162138 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:53:00 crc kubenswrapper[4787]: I1001 09:53:00.167144 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0-internal-tls-certs\") pod \"neutron-58d9cd74f7-86sts\" (UID: \"1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0\") " pod="openstack/neutron-58d9cd74f7-86sts" Oct 01 09:53:00 crc kubenswrapper[4787]: W1001 09:53:00.181183 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb4ccc778_bb43_4e30_b9e2_01fe53ba7f3c.slice/crio-e3cc61e669df8493afec070a71fc990a8c3efae859954713f94e0f84e58643af WatchSource:0}: Error finding container e3cc61e669df8493afec070a71fc990a8c3efae859954713f94e0f84e58643af: Status 404 returned error can't find the container with id e3cc61e669df8493afec070a71fc990a8c3efae859954713f94e0f84e58643af Oct 01 09:53:00 crc kubenswrapper[4787]: I1001 09:53:00.245703 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-58d9cd74f7-86sts" Oct 01 09:53:00 crc kubenswrapper[4787]: I1001 09:53:00.534245 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9587d64e-33e3-45d2-8ee7-a776fcc60d88" path="/var/lib/kubelet/pods/9587d64e-33e3-45d2-8ee7-a776fcc60d88/volumes" Oct 01 09:53:00 crc kubenswrapper[4787]: I1001 09:53:00.799403 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-58d9cd74f7-86sts"] Oct 01 09:53:00 crc kubenswrapper[4787]: W1001 09:53:00.813736 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1e4d70a9_10b1_4bb7_a151_56f16bb2cbe0.slice/crio-2e9c3758692c4ae55625907a3333854b707b58b2cd6faa765550c40c67b35bfa WatchSource:0}: Error finding container 2e9c3758692c4ae55625907a3333854b707b58b2cd6faa765550c40c67b35bfa: Status 404 returned error can't find the container with id 2e9c3758692c4ae55625907a3333854b707b58b2cd6faa765550c40c67b35bfa Oct 01 09:53:01 crc kubenswrapper[4787]: I1001 09:53:01.155790 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-58d9cd74f7-86sts" event={"ID":"1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0","Type":"ContainerStarted","Data":"a600fee301175850d09afd420891cb54bccf8bcb09e1382ac067239f3f2f7bde"} Oct 01 09:53:01 crc kubenswrapper[4787]: I1001 09:53:01.156318 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-58d9cd74f7-86sts" event={"ID":"1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0","Type":"ContainerStarted","Data":"2e9c3758692c4ae55625907a3333854b707b58b2cd6faa765550c40c67b35bfa"} Oct 01 09:53:01 crc kubenswrapper[4787]: I1001 09:53:01.158637 4787 generic.go:334] "Generic (PLEG): container finished" podID="7983e213-c317-42b2-8944-85bdb78dac8f" containerID="c363923f03c8431731ce61fb7088f6a51538ab7d45b2f5fb59d3d2d245ae02f0" exitCode=0 Oct 01 09:53:01 crc kubenswrapper[4787]: I1001 09:53:01.158710 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67879968b-d5598" event={"ID":"7983e213-c317-42b2-8944-85bdb78dac8f","Type":"ContainerDied","Data":"c363923f03c8431731ce61fb7088f6a51538ab7d45b2f5fb59d3d2d245ae02f0"} Oct 01 09:53:01 crc kubenswrapper[4787]: I1001 09:53:01.160522 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c","Type":"ContainerStarted","Data":"98c33f6c04008cd30bdef34a1de003ea58ef86bd06cb44398a6e45d9e7916bc5"} Oct 01 09:53:01 crc kubenswrapper[4787]: I1001 09:53:01.160562 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c","Type":"ContainerStarted","Data":"e3cc61e669df8493afec070a71fc990a8c3efae859954713f94e0f84e58643af"} Oct 01 09:53:01 crc kubenswrapper[4787]: I1001 09:53:01.642828 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-wndd6" Oct 01 09:53:01 crc kubenswrapper[4787]: I1001 09:53:01.758677 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/15111beb-b05b-440f-a4c9-077eca1c37d1-db-sync-config-data\") pod \"15111beb-b05b-440f-a4c9-077eca1c37d1\" (UID: \"15111beb-b05b-440f-a4c9-077eca1c37d1\") " Oct 01 09:53:01 crc kubenswrapper[4787]: I1001 09:53:01.758769 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15111beb-b05b-440f-a4c9-077eca1c37d1-combined-ca-bundle\") pod \"15111beb-b05b-440f-a4c9-077eca1c37d1\" (UID: \"15111beb-b05b-440f-a4c9-077eca1c37d1\") " Oct 01 09:53:01 crc kubenswrapper[4787]: I1001 09:53:01.758839 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nlgx\" (UniqueName: \"kubernetes.io/projected/15111beb-b05b-440f-a4c9-077eca1c37d1-kube-api-access-8nlgx\") pod \"15111beb-b05b-440f-a4c9-077eca1c37d1\" (UID: \"15111beb-b05b-440f-a4c9-077eca1c37d1\") " Oct 01 09:53:01 crc kubenswrapper[4787]: I1001 09:53:01.763391 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15111beb-b05b-440f-a4c9-077eca1c37d1-kube-api-access-8nlgx" (OuterVolumeSpecName: "kube-api-access-8nlgx") pod "15111beb-b05b-440f-a4c9-077eca1c37d1" (UID: "15111beb-b05b-440f-a4c9-077eca1c37d1"). InnerVolumeSpecName "kube-api-access-8nlgx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:53:01 crc kubenswrapper[4787]: I1001 09:53:01.764410 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15111beb-b05b-440f-a4c9-077eca1c37d1-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "15111beb-b05b-440f-a4c9-077eca1c37d1" (UID: "15111beb-b05b-440f-a4c9-077eca1c37d1"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:01 crc kubenswrapper[4787]: I1001 09:53:01.788493 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15111beb-b05b-440f-a4c9-077eca1c37d1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "15111beb-b05b-440f-a4c9-077eca1c37d1" (UID: "15111beb-b05b-440f-a4c9-077eca1c37d1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:01 crc kubenswrapper[4787]: I1001 09:53:01.860906 4787 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/15111beb-b05b-440f-a4c9-077eca1c37d1-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:01 crc kubenswrapper[4787]: I1001 09:53:01.860945 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15111beb-b05b-440f-a4c9-077eca1c37d1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:01 crc kubenswrapper[4787]: I1001 09:53:01.860958 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nlgx\" (UniqueName: \"kubernetes.io/projected/15111beb-b05b-440f-a4c9-077eca1c37d1-kube-api-access-8nlgx\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.199318 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c","Type":"ContainerStarted","Data":"974cfa14c1d8f423c33e89b90de41d56b73acf24136c57d8454283e7000a8425"} Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.203033 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-wndd6" event={"ID":"15111beb-b05b-440f-a4c9-077eca1c37d1","Type":"ContainerDied","Data":"7d820565c6008eacb9e6bdcd4e423c7a4bb1a7eebd7e955b253647975da4e182"} Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.203093 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d820565c6008eacb9e6bdcd4e423c7a4bb1a7eebd7e955b253647975da4e182" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.203187 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-wndd6" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.226124 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-58d9cd74f7-86sts" event={"ID":"1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0","Type":"ContainerStarted","Data":"1b6d559fa2750154902e93ac0a3be35777f57946fef6910f6c36aaa1fd2a3463"} Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.227814 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-58d9cd74f7-86sts" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.262213 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-58d9cd74f7-86sts" podStartSLOduration=3.262181538 podStartE2EDuration="3.262181538s" podCreationTimestamp="2025-10-01 09:52:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:53:02.252179265 +0000 UTC m=+1014.367323422" watchObservedRunningTime="2025-10-01 09:53:02.262181538 +0000 UTC m=+1014.377325695" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.464551 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-64d9c8b8f9-zv7w7"] Oct 01 09:53:02 crc kubenswrapper[4787]: E1001 09:53:02.465160 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15111beb-b05b-440f-a4c9-077eca1c37d1" containerName="barbican-db-sync" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.465180 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="15111beb-b05b-440f-a4c9-077eca1c37d1" containerName="barbican-db-sync" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.465403 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="15111beb-b05b-440f-a4c9-077eca1c37d1" containerName="barbican-db-sync" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.466763 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-64d9c8b8f9-zv7w7" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.472731 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.473240 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.473602 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-6h4ql" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.488061 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-699b4d5c4-jdg79"] Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.490334 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-699b4d5c4-jdg79" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.500360 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.518710 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-64d9c8b8f9-zv7w7"] Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.554007 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-699b4d5c4-jdg79"] Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.577848 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zk7xl\" (UniqueName: \"kubernetes.io/projected/484a5c7d-7633-4e34-a235-af316bbccada-kube-api-access-zk7xl\") pod \"barbican-keystone-listener-699b4d5c4-jdg79\" (UID: \"484a5c7d-7633-4e34-a235-af316bbccada\") " pod="openstack/barbican-keystone-listener-699b4d5c4-jdg79" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.577953 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea69f2cb-9d4f-4b3c-a0ba-8a3fc8fce274-logs\") pod \"barbican-worker-64d9c8b8f9-zv7w7\" (UID: \"ea69f2cb-9d4f-4b3c-a0ba-8a3fc8fce274\") " pod="openstack/barbican-worker-64d9c8b8f9-zv7w7" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.577997 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/484a5c7d-7633-4e34-a235-af316bbccada-config-data\") pod \"barbican-keystone-listener-699b4d5c4-jdg79\" (UID: \"484a5c7d-7633-4e34-a235-af316bbccada\") " pod="openstack/barbican-keystone-listener-699b4d5c4-jdg79" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.578104 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ea69f2cb-9d4f-4b3c-a0ba-8a3fc8fce274-config-data-custom\") pod \"barbican-worker-64d9c8b8f9-zv7w7\" (UID: \"ea69f2cb-9d4f-4b3c-a0ba-8a3fc8fce274\") " pod="openstack/barbican-worker-64d9c8b8f9-zv7w7" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.578180 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fj8vf\" (UniqueName: \"kubernetes.io/projected/ea69f2cb-9d4f-4b3c-a0ba-8a3fc8fce274-kube-api-access-fj8vf\") pod \"barbican-worker-64d9c8b8f9-zv7w7\" (UID: \"ea69f2cb-9d4f-4b3c-a0ba-8a3fc8fce274\") " pod="openstack/barbican-worker-64d9c8b8f9-zv7w7" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.578227 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea69f2cb-9d4f-4b3c-a0ba-8a3fc8fce274-combined-ca-bundle\") pod \"barbican-worker-64d9c8b8f9-zv7w7\" (UID: \"ea69f2cb-9d4f-4b3c-a0ba-8a3fc8fce274\") " pod="openstack/barbican-worker-64d9c8b8f9-zv7w7" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.578252 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/484a5c7d-7633-4e34-a235-af316bbccada-combined-ca-bundle\") pod \"barbican-keystone-listener-699b4d5c4-jdg79\" (UID: \"484a5c7d-7633-4e34-a235-af316bbccada\") " pod="openstack/barbican-keystone-listener-699b4d5c4-jdg79" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.578281 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/484a5c7d-7633-4e34-a235-af316bbccada-logs\") pod \"barbican-keystone-listener-699b4d5c4-jdg79\" (UID: \"484a5c7d-7633-4e34-a235-af316bbccada\") " pod="openstack/barbican-keystone-listener-699b4d5c4-jdg79" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.578346 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/484a5c7d-7633-4e34-a235-af316bbccada-config-data-custom\") pod \"barbican-keystone-listener-699b4d5c4-jdg79\" (UID: \"484a5c7d-7633-4e34-a235-af316bbccada\") " pod="openstack/barbican-keystone-listener-699b4d5c4-jdg79" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.578391 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea69f2cb-9d4f-4b3c-a0ba-8a3fc8fce274-config-data\") pod \"barbican-worker-64d9c8b8f9-zv7w7\" (UID: \"ea69f2cb-9d4f-4b3c-a0ba-8a3fc8fce274\") " pod="openstack/barbican-worker-64d9c8b8f9-zv7w7" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.609660 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-c4db58965-h4fwp"] Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.609932 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-c4db58965-h4fwp" podUID="a7835443-3da0-4c3f-8643-b07cbf790a17" containerName="dnsmasq-dns" containerID="cri-o://f604cceb35751cdf4ced9c10d851eef06a4de09e5f2c818aa1f04a04d70ef53c" gracePeriod=10 Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.642306 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f4d8965b5-7dw2j"] Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.644341 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f4d8965b5-7dw2j" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.682056 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/484a5c7d-7633-4e34-a235-af316bbccada-combined-ca-bundle\") pod \"barbican-keystone-listener-699b4d5c4-jdg79\" (UID: \"484a5c7d-7633-4e34-a235-af316bbccada\") " pod="openstack/barbican-keystone-listener-699b4d5c4-jdg79" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.682214 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/484a5c7d-7633-4e34-a235-af316bbccada-logs\") pod \"barbican-keystone-listener-699b4d5c4-jdg79\" (UID: \"484a5c7d-7633-4e34-a235-af316bbccada\") " pod="openstack/barbican-keystone-listener-699b4d5c4-jdg79" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.683285 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjgm2\" (UniqueName: \"kubernetes.io/projected/17fa3c34-c304-4972-b023-879f3ee28416-kube-api-access-sjgm2\") pod \"dnsmasq-dns-f4d8965b5-7dw2j\" (UID: \"17fa3c34-c304-4972-b023-879f3ee28416\") " pod="openstack/dnsmasq-dns-f4d8965b5-7dw2j" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.683366 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f4d8965b5-7dw2j"] Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.683379 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/484a5c7d-7633-4e34-a235-af316bbccada-config-data-custom\") pod \"barbican-keystone-listener-699b4d5c4-jdg79\" (UID: \"484a5c7d-7633-4e34-a235-af316bbccada\") " pod="openstack/barbican-keystone-listener-699b4d5c4-jdg79" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.683544 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea69f2cb-9d4f-4b3c-a0ba-8a3fc8fce274-config-data\") pod \"barbican-worker-64d9c8b8f9-zv7w7\" (UID: \"ea69f2cb-9d4f-4b3c-a0ba-8a3fc8fce274\") " pod="openstack/barbican-worker-64d9c8b8f9-zv7w7" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.683709 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zk7xl\" (UniqueName: \"kubernetes.io/projected/484a5c7d-7633-4e34-a235-af316bbccada-kube-api-access-zk7xl\") pod \"barbican-keystone-listener-699b4d5c4-jdg79\" (UID: \"484a5c7d-7633-4e34-a235-af316bbccada\") " pod="openstack/barbican-keystone-listener-699b4d5c4-jdg79" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.683745 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17fa3c34-c304-4972-b023-879f3ee28416-ovsdbserver-nb\") pod \"dnsmasq-dns-f4d8965b5-7dw2j\" (UID: \"17fa3c34-c304-4972-b023-879f3ee28416\") " pod="openstack/dnsmasq-dns-f4d8965b5-7dw2j" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.683781 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/17fa3c34-c304-4972-b023-879f3ee28416-dns-swift-storage-0\") pod \"dnsmasq-dns-f4d8965b5-7dw2j\" (UID: \"17fa3c34-c304-4972-b023-879f3ee28416\") " pod="openstack/dnsmasq-dns-f4d8965b5-7dw2j" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.683813 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17fa3c34-c304-4972-b023-879f3ee28416-ovsdbserver-sb\") pod \"dnsmasq-dns-f4d8965b5-7dw2j\" (UID: \"17fa3c34-c304-4972-b023-879f3ee28416\") " pod="openstack/dnsmasq-dns-f4d8965b5-7dw2j" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.683845 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea69f2cb-9d4f-4b3c-a0ba-8a3fc8fce274-logs\") pod \"barbican-worker-64d9c8b8f9-zv7w7\" (UID: \"ea69f2cb-9d4f-4b3c-a0ba-8a3fc8fce274\") " pod="openstack/barbican-worker-64d9c8b8f9-zv7w7" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.683993 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/484a5c7d-7633-4e34-a235-af316bbccada-config-data\") pod \"barbican-keystone-listener-699b4d5c4-jdg79\" (UID: \"484a5c7d-7633-4e34-a235-af316bbccada\") " pod="openstack/barbican-keystone-listener-699b4d5c4-jdg79" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.684189 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ea69f2cb-9d4f-4b3c-a0ba-8a3fc8fce274-config-data-custom\") pod \"barbican-worker-64d9c8b8f9-zv7w7\" (UID: \"ea69f2cb-9d4f-4b3c-a0ba-8a3fc8fce274\") " pod="openstack/barbican-worker-64d9c8b8f9-zv7w7" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.684222 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17fa3c34-c304-4972-b023-879f3ee28416-config\") pod \"dnsmasq-dns-f4d8965b5-7dw2j\" (UID: \"17fa3c34-c304-4972-b023-879f3ee28416\") " pod="openstack/dnsmasq-dns-f4d8965b5-7dw2j" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.684282 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17fa3c34-c304-4972-b023-879f3ee28416-dns-svc\") pod \"dnsmasq-dns-f4d8965b5-7dw2j\" (UID: \"17fa3c34-c304-4972-b023-879f3ee28416\") " pod="openstack/dnsmasq-dns-f4d8965b5-7dw2j" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.684354 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea69f2cb-9d4f-4b3c-a0ba-8a3fc8fce274-logs\") pod \"barbican-worker-64d9c8b8f9-zv7w7\" (UID: \"ea69f2cb-9d4f-4b3c-a0ba-8a3fc8fce274\") " pod="openstack/barbican-worker-64d9c8b8f9-zv7w7" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.684391 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fj8vf\" (UniqueName: \"kubernetes.io/projected/ea69f2cb-9d4f-4b3c-a0ba-8a3fc8fce274-kube-api-access-fj8vf\") pod \"barbican-worker-64d9c8b8f9-zv7w7\" (UID: \"ea69f2cb-9d4f-4b3c-a0ba-8a3fc8fce274\") " pod="openstack/barbican-worker-64d9c8b8f9-zv7w7" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.684437 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea69f2cb-9d4f-4b3c-a0ba-8a3fc8fce274-combined-ca-bundle\") pod \"barbican-worker-64d9c8b8f9-zv7w7\" (UID: \"ea69f2cb-9d4f-4b3c-a0ba-8a3fc8fce274\") " pod="openstack/barbican-worker-64d9c8b8f9-zv7w7" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.684588 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/484a5c7d-7633-4e34-a235-af316bbccada-logs\") pod \"barbican-keystone-listener-699b4d5c4-jdg79\" (UID: \"484a5c7d-7633-4e34-a235-af316bbccada\") " pod="openstack/barbican-keystone-listener-699b4d5c4-jdg79" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.696056 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea69f2cb-9d4f-4b3c-a0ba-8a3fc8fce274-combined-ca-bundle\") pod \"barbican-worker-64d9c8b8f9-zv7w7\" (UID: \"ea69f2cb-9d4f-4b3c-a0ba-8a3fc8fce274\") " pod="openstack/barbican-worker-64d9c8b8f9-zv7w7" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.707932 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea69f2cb-9d4f-4b3c-a0ba-8a3fc8fce274-config-data\") pod \"barbican-worker-64d9c8b8f9-zv7w7\" (UID: \"ea69f2cb-9d4f-4b3c-a0ba-8a3fc8fce274\") " pod="openstack/barbican-worker-64d9c8b8f9-zv7w7" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.708591 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/484a5c7d-7633-4e34-a235-af316bbccada-config-data-custom\") pod \"barbican-keystone-listener-699b4d5c4-jdg79\" (UID: \"484a5c7d-7633-4e34-a235-af316bbccada\") " pod="openstack/barbican-keystone-listener-699b4d5c4-jdg79" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.710642 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/484a5c7d-7633-4e34-a235-af316bbccada-config-data\") pod \"barbican-keystone-listener-699b4d5c4-jdg79\" (UID: \"484a5c7d-7633-4e34-a235-af316bbccada\") " pod="openstack/barbican-keystone-listener-699b4d5c4-jdg79" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.715342 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ea69f2cb-9d4f-4b3c-a0ba-8a3fc8fce274-config-data-custom\") pod \"barbican-worker-64d9c8b8f9-zv7w7\" (UID: \"ea69f2cb-9d4f-4b3c-a0ba-8a3fc8fce274\") " pod="openstack/barbican-worker-64d9c8b8f9-zv7w7" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.741190 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-79b69df69b-mfcjt"] Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.743770 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-79b69df69b-mfcjt" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.749552 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.761231 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/484a5c7d-7633-4e34-a235-af316bbccada-combined-ca-bundle\") pod \"barbican-keystone-listener-699b4d5c4-jdg79\" (UID: \"484a5c7d-7633-4e34-a235-af316bbccada\") " pod="openstack/barbican-keystone-listener-699b4d5c4-jdg79" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.761338 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-79b69df69b-mfcjt"] Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.761737 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fj8vf\" (UniqueName: \"kubernetes.io/projected/ea69f2cb-9d4f-4b3c-a0ba-8a3fc8fce274-kube-api-access-fj8vf\") pod \"barbican-worker-64d9c8b8f9-zv7w7\" (UID: \"ea69f2cb-9d4f-4b3c-a0ba-8a3fc8fce274\") " pod="openstack/barbican-worker-64d9c8b8f9-zv7w7" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.785905 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0244e33a-702c-412d-8074-a2c910f375d0-logs\") pod \"barbican-api-79b69df69b-mfcjt\" (UID: \"0244e33a-702c-412d-8074-a2c910f375d0\") " pod="openstack/barbican-api-79b69df69b-mfcjt" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.785985 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17fa3c34-c304-4972-b023-879f3ee28416-ovsdbserver-nb\") pod \"dnsmasq-dns-f4d8965b5-7dw2j\" (UID: \"17fa3c34-c304-4972-b023-879f3ee28416\") " pod="openstack/dnsmasq-dns-f4d8965b5-7dw2j" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.786012 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/17fa3c34-c304-4972-b023-879f3ee28416-dns-swift-storage-0\") pod \"dnsmasq-dns-f4d8965b5-7dw2j\" (UID: \"17fa3c34-c304-4972-b023-879f3ee28416\") " pod="openstack/dnsmasq-dns-f4d8965b5-7dw2j" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.786034 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17fa3c34-c304-4972-b023-879f3ee28416-ovsdbserver-sb\") pod \"dnsmasq-dns-f4d8965b5-7dw2j\" (UID: \"17fa3c34-c304-4972-b023-879f3ee28416\") " pod="openstack/dnsmasq-dns-f4d8965b5-7dw2j" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.786069 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0244e33a-702c-412d-8074-a2c910f375d0-combined-ca-bundle\") pod \"barbican-api-79b69df69b-mfcjt\" (UID: \"0244e33a-702c-412d-8074-a2c910f375d0\") " pod="openstack/barbican-api-79b69df69b-mfcjt" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.786119 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0244e33a-702c-412d-8074-a2c910f375d0-config-data\") pod \"barbican-api-79b69df69b-mfcjt\" (UID: \"0244e33a-702c-412d-8074-a2c910f375d0\") " pod="openstack/barbican-api-79b69df69b-mfcjt" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.786265 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdw8q\" (UniqueName: \"kubernetes.io/projected/0244e33a-702c-412d-8074-a2c910f375d0-kube-api-access-kdw8q\") pod \"barbican-api-79b69df69b-mfcjt\" (UID: \"0244e33a-702c-412d-8074-a2c910f375d0\") " pod="openstack/barbican-api-79b69df69b-mfcjt" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.786321 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17fa3c34-c304-4972-b023-879f3ee28416-config\") pod \"dnsmasq-dns-f4d8965b5-7dw2j\" (UID: \"17fa3c34-c304-4972-b023-879f3ee28416\") " pod="openstack/dnsmasq-dns-f4d8965b5-7dw2j" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.786347 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0244e33a-702c-412d-8074-a2c910f375d0-config-data-custom\") pod \"barbican-api-79b69df69b-mfcjt\" (UID: \"0244e33a-702c-412d-8074-a2c910f375d0\") " pod="openstack/barbican-api-79b69df69b-mfcjt" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.786382 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17fa3c34-c304-4972-b023-879f3ee28416-dns-svc\") pod \"dnsmasq-dns-f4d8965b5-7dw2j\" (UID: \"17fa3c34-c304-4972-b023-879f3ee28416\") " pod="openstack/dnsmasq-dns-f4d8965b5-7dw2j" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.786466 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjgm2\" (UniqueName: \"kubernetes.io/projected/17fa3c34-c304-4972-b023-879f3ee28416-kube-api-access-sjgm2\") pod \"dnsmasq-dns-f4d8965b5-7dw2j\" (UID: \"17fa3c34-c304-4972-b023-879f3ee28416\") " pod="openstack/dnsmasq-dns-f4d8965b5-7dw2j" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.797415 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-64d9c8b8f9-zv7w7" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.800278 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zk7xl\" (UniqueName: \"kubernetes.io/projected/484a5c7d-7633-4e34-a235-af316bbccada-kube-api-access-zk7xl\") pod \"barbican-keystone-listener-699b4d5c4-jdg79\" (UID: \"484a5c7d-7633-4e34-a235-af316bbccada\") " pod="openstack/barbican-keystone-listener-699b4d5c4-jdg79" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.809759 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17fa3c34-c304-4972-b023-879f3ee28416-ovsdbserver-nb\") pod \"dnsmasq-dns-f4d8965b5-7dw2j\" (UID: \"17fa3c34-c304-4972-b023-879f3ee28416\") " pod="openstack/dnsmasq-dns-f4d8965b5-7dw2j" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.809841 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17fa3c34-c304-4972-b023-879f3ee28416-ovsdbserver-sb\") pod \"dnsmasq-dns-f4d8965b5-7dw2j\" (UID: \"17fa3c34-c304-4972-b023-879f3ee28416\") " pod="openstack/dnsmasq-dns-f4d8965b5-7dw2j" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.810456 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/17fa3c34-c304-4972-b023-879f3ee28416-dns-swift-storage-0\") pod \"dnsmasq-dns-f4d8965b5-7dw2j\" (UID: \"17fa3c34-c304-4972-b023-879f3ee28416\") " pod="openstack/dnsmasq-dns-f4d8965b5-7dw2j" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.810981 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17fa3c34-c304-4972-b023-879f3ee28416-config\") pod \"dnsmasq-dns-f4d8965b5-7dw2j\" (UID: \"17fa3c34-c304-4972-b023-879f3ee28416\") " pod="openstack/dnsmasq-dns-f4d8965b5-7dw2j" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.812004 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17fa3c34-c304-4972-b023-879f3ee28416-dns-svc\") pod \"dnsmasq-dns-f4d8965b5-7dw2j\" (UID: \"17fa3c34-c304-4972-b023-879f3ee28416\") " pod="openstack/dnsmasq-dns-f4d8965b5-7dw2j" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.820652 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-699b4d5c4-jdg79" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.826567 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjgm2\" (UniqueName: \"kubernetes.io/projected/17fa3c34-c304-4972-b023-879f3ee28416-kube-api-access-sjgm2\") pod \"dnsmasq-dns-f4d8965b5-7dw2j\" (UID: \"17fa3c34-c304-4972-b023-879f3ee28416\") " pod="openstack/dnsmasq-dns-f4d8965b5-7dw2j" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.900572 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0244e33a-702c-412d-8074-a2c910f375d0-config-data-custom\") pod \"barbican-api-79b69df69b-mfcjt\" (UID: \"0244e33a-702c-412d-8074-a2c910f375d0\") " pod="openstack/barbican-api-79b69df69b-mfcjt" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.900715 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0244e33a-702c-412d-8074-a2c910f375d0-logs\") pod \"barbican-api-79b69df69b-mfcjt\" (UID: \"0244e33a-702c-412d-8074-a2c910f375d0\") " pod="openstack/barbican-api-79b69df69b-mfcjt" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.900783 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0244e33a-702c-412d-8074-a2c910f375d0-combined-ca-bundle\") pod \"barbican-api-79b69df69b-mfcjt\" (UID: \"0244e33a-702c-412d-8074-a2c910f375d0\") " pod="openstack/barbican-api-79b69df69b-mfcjt" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.900803 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0244e33a-702c-412d-8074-a2c910f375d0-config-data\") pod \"barbican-api-79b69df69b-mfcjt\" (UID: \"0244e33a-702c-412d-8074-a2c910f375d0\") " pod="openstack/barbican-api-79b69df69b-mfcjt" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.900832 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdw8q\" (UniqueName: \"kubernetes.io/projected/0244e33a-702c-412d-8074-a2c910f375d0-kube-api-access-kdw8q\") pod \"barbican-api-79b69df69b-mfcjt\" (UID: \"0244e33a-702c-412d-8074-a2c910f375d0\") " pod="openstack/barbican-api-79b69df69b-mfcjt" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.901395 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0244e33a-702c-412d-8074-a2c910f375d0-logs\") pod \"barbican-api-79b69df69b-mfcjt\" (UID: \"0244e33a-702c-412d-8074-a2c910f375d0\") " pod="openstack/barbican-api-79b69df69b-mfcjt" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.916960 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0244e33a-702c-412d-8074-a2c910f375d0-combined-ca-bundle\") pod \"barbican-api-79b69df69b-mfcjt\" (UID: \"0244e33a-702c-412d-8074-a2c910f375d0\") " pod="openstack/barbican-api-79b69df69b-mfcjt" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.917302 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0244e33a-702c-412d-8074-a2c910f375d0-config-data-custom\") pod \"barbican-api-79b69df69b-mfcjt\" (UID: \"0244e33a-702c-412d-8074-a2c910f375d0\") " pod="openstack/barbican-api-79b69df69b-mfcjt" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.918752 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0244e33a-702c-412d-8074-a2c910f375d0-config-data\") pod \"barbican-api-79b69df69b-mfcjt\" (UID: \"0244e33a-702c-412d-8074-a2c910f375d0\") " pod="openstack/barbican-api-79b69df69b-mfcjt" Oct 01 09:53:02 crc kubenswrapper[4787]: I1001 09:53:02.920198 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdw8q\" (UniqueName: \"kubernetes.io/projected/0244e33a-702c-412d-8074-a2c910f375d0-kube-api-access-kdw8q\") pod \"barbican-api-79b69df69b-mfcjt\" (UID: \"0244e33a-702c-412d-8074-a2c910f375d0\") " pod="openstack/barbican-api-79b69df69b-mfcjt" Oct 01 09:53:03 crc kubenswrapper[4787]: I1001 09:53:03.124365 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f4d8965b5-7dw2j" Oct 01 09:53:03 crc kubenswrapper[4787]: I1001 09:53:03.138517 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-79b69df69b-mfcjt" Oct 01 09:53:03 crc kubenswrapper[4787]: I1001 09:53:03.176749 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-67879968b-d5598" podUID="7983e213-c317-42b2-8944-85bdb78dac8f" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Oct 01 09:53:03 crc kubenswrapper[4787]: I1001 09:53:03.241881 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c","Type":"ContainerStarted","Data":"9ea5042ebd4a4cfdce6dd4dd8b3f130111459a9b369f9c3d2aa1cb0f72d20ad0"} Oct 01 09:53:03 crc kubenswrapper[4787]: I1001 09:53:03.243858 4787 generic.go:334] "Generic (PLEG): container finished" podID="a7835443-3da0-4c3f-8643-b07cbf790a17" containerID="f604cceb35751cdf4ced9c10d851eef06a4de09e5f2c818aa1f04a04d70ef53c" exitCode=0 Oct 01 09:53:03 crc kubenswrapper[4787]: I1001 09:53:03.243896 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c4db58965-h4fwp" event={"ID":"a7835443-3da0-4c3f-8643-b07cbf790a17","Type":"ContainerDied","Data":"f604cceb35751cdf4ced9c10d851eef06a4de09e5f2c818aa1f04a04d70ef53c"} Oct 01 09:53:03 crc kubenswrapper[4787]: I1001 09:53:03.243925 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c4db58965-h4fwp" event={"ID":"a7835443-3da0-4c3f-8643-b07cbf790a17","Type":"ContainerDied","Data":"c1bc7b380d40fc56ea801fca38d39015451179ce2f681cd59fdd6014a7d8f74e"} Oct 01 09:53:03 crc kubenswrapper[4787]: I1001 09:53:03.243937 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1bc7b380d40fc56ea801fca38d39015451179ce2f681cd59fdd6014a7d8f74e" Oct 01 09:53:03 crc kubenswrapper[4787]: I1001 09:53:03.413341 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-64d9c8b8f9-zv7w7"] Oct 01 09:53:03 crc kubenswrapper[4787]: I1001 09:53:03.420700 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c4db58965-h4fwp" Oct 01 09:53:03 crc kubenswrapper[4787]: I1001 09:53:03.430796 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-699b4d5c4-jdg79"] Oct 01 09:53:03 crc kubenswrapper[4787]: W1001 09:53:03.482328 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod484a5c7d_7633_4e34_a235_af316bbccada.slice/crio-4e5025b17bff32fc628c55fb38ff4519d21ab89d11e5388d8c4a6cb4392ab891 WatchSource:0}: Error finding container 4e5025b17bff32fc628c55fb38ff4519d21ab89d11e5388d8c4a6cb4392ab891: Status 404 returned error can't find the container with id 4e5025b17bff32fc628c55fb38ff4519d21ab89d11e5388d8c4a6cb4392ab891 Oct 01 09:53:03 crc kubenswrapper[4787]: I1001 09:53:03.632544 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49kwv\" (UniqueName: \"kubernetes.io/projected/a7835443-3da0-4c3f-8643-b07cbf790a17-kube-api-access-49kwv\") pod \"a7835443-3da0-4c3f-8643-b07cbf790a17\" (UID: \"a7835443-3da0-4c3f-8643-b07cbf790a17\") " Oct 01 09:53:03 crc kubenswrapper[4787]: I1001 09:53:03.632641 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7835443-3da0-4c3f-8643-b07cbf790a17-dns-svc\") pod \"a7835443-3da0-4c3f-8643-b07cbf790a17\" (UID: \"a7835443-3da0-4c3f-8643-b07cbf790a17\") " Oct 01 09:53:03 crc kubenswrapper[4787]: I1001 09:53:03.632665 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7835443-3da0-4c3f-8643-b07cbf790a17-ovsdbserver-nb\") pod \"a7835443-3da0-4c3f-8643-b07cbf790a17\" (UID: \"a7835443-3da0-4c3f-8643-b07cbf790a17\") " Oct 01 09:53:03 crc kubenswrapper[4787]: I1001 09:53:03.632696 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7835443-3da0-4c3f-8643-b07cbf790a17-config\") pod \"a7835443-3da0-4c3f-8643-b07cbf790a17\" (UID: \"a7835443-3da0-4c3f-8643-b07cbf790a17\") " Oct 01 09:53:03 crc kubenswrapper[4787]: I1001 09:53:03.632728 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a7835443-3da0-4c3f-8643-b07cbf790a17-dns-swift-storage-0\") pod \"a7835443-3da0-4c3f-8643-b07cbf790a17\" (UID: \"a7835443-3da0-4c3f-8643-b07cbf790a17\") " Oct 01 09:53:03 crc kubenswrapper[4787]: I1001 09:53:03.632776 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7835443-3da0-4c3f-8643-b07cbf790a17-ovsdbserver-sb\") pod \"a7835443-3da0-4c3f-8643-b07cbf790a17\" (UID: \"a7835443-3da0-4c3f-8643-b07cbf790a17\") " Oct 01 09:53:03 crc kubenswrapper[4787]: I1001 09:53:03.641401 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7835443-3da0-4c3f-8643-b07cbf790a17-kube-api-access-49kwv" (OuterVolumeSpecName: "kube-api-access-49kwv") pod "a7835443-3da0-4c3f-8643-b07cbf790a17" (UID: "a7835443-3da0-4c3f-8643-b07cbf790a17"). InnerVolumeSpecName "kube-api-access-49kwv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:53:03 crc kubenswrapper[4787]: I1001 09:53:03.718014 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7835443-3da0-4c3f-8643-b07cbf790a17-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a7835443-3da0-4c3f-8643-b07cbf790a17" (UID: "a7835443-3da0-4c3f-8643-b07cbf790a17"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:53:03 crc kubenswrapper[4787]: I1001 09:53:03.733670 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7835443-3da0-4c3f-8643-b07cbf790a17-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a7835443-3da0-4c3f-8643-b07cbf790a17" (UID: "a7835443-3da0-4c3f-8643-b07cbf790a17"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:53:03 crc kubenswrapper[4787]: I1001 09:53:03.734164 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7835443-3da0-4c3f-8643-b07cbf790a17-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a7835443-3da0-4c3f-8643-b07cbf790a17" (UID: "a7835443-3da0-4c3f-8643-b07cbf790a17"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:53:03 crc kubenswrapper[4787]: I1001 09:53:03.734391 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7835443-3da0-4c3f-8643-b07cbf790a17-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a7835443-3da0-4c3f-8643-b07cbf790a17" (UID: "a7835443-3da0-4c3f-8643-b07cbf790a17"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:53:03 crc kubenswrapper[4787]: I1001 09:53:03.734679 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7835443-3da0-4c3f-8643-b07cbf790a17-dns-svc\") pod \"a7835443-3da0-4c3f-8643-b07cbf790a17\" (UID: \"a7835443-3da0-4c3f-8643-b07cbf790a17\") " Oct 01 09:53:03 crc kubenswrapper[4787]: W1001 09:53:03.734931 4787 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/a7835443-3da0-4c3f-8643-b07cbf790a17/volumes/kubernetes.io~configmap/dns-svc Oct 01 09:53:03 crc kubenswrapper[4787]: I1001 09:53:03.734945 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7835443-3da0-4c3f-8643-b07cbf790a17-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a7835443-3da0-4c3f-8643-b07cbf790a17" (UID: "a7835443-3da0-4c3f-8643-b07cbf790a17"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:53:03 crc kubenswrapper[4787]: I1001 09:53:03.735589 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7835443-3da0-4c3f-8643-b07cbf790a17-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:03 crc kubenswrapper[4787]: I1001 09:53:03.735611 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49kwv\" (UniqueName: \"kubernetes.io/projected/a7835443-3da0-4c3f-8643-b07cbf790a17-kube-api-access-49kwv\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:03 crc kubenswrapper[4787]: I1001 09:53:03.735624 4787 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7835443-3da0-4c3f-8643-b07cbf790a17-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:03 crc kubenswrapper[4787]: I1001 09:53:03.735640 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7835443-3da0-4c3f-8643-b07cbf790a17-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:03 crc kubenswrapper[4787]: I1001 09:53:03.735648 4787 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a7835443-3da0-4c3f-8643-b07cbf790a17-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:03 crc kubenswrapper[4787]: I1001 09:53:03.737508 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7835443-3da0-4c3f-8643-b07cbf790a17-config" (OuterVolumeSpecName: "config") pod "a7835443-3da0-4c3f-8643-b07cbf790a17" (UID: "a7835443-3da0-4c3f-8643-b07cbf790a17"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:53:03 crc kubenswrapper[4787]: I1001 09:53:03.780038 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f4d8965b5-7dw2j"] Oct 01 09:53:03 crc kubenswrapper[4787]: I1001 09:53:03.841354 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7835443-3da0-4c3f-8643-b07cbf790a17-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:03 crc kubenswrapper[4787]: I1001 09:53:03.932086 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-79b69df69b-mfcjt"] Oct 01 09:53:04 crc kubenswrapper[4787]: I1001 09:53:04.260374 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c","Type":"ContainerStarted","Data":"ae68e6f7fc578d6c34f91b087188337008568de65db874e1e8a7966190fdc84e"} Oct 01 09:53:04 crc kubenswrapper[4787]: I1001 09:53:04.261811 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 01 09:53:04 crc kubenswrapper[4787]: I1001 09:53:04.273606 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-64d9c8b8f9-zv7w7" event={"ID":"ea69f2cb-9d4f-4b3c-a0ba-8a3fc8fce274","Type":"ContainerStarted","Data":"834bc6113226a775c727e4bdb6ed737d2b9b82f4730bc59811653ae94c63e25c"} Oct 01 09:53:04 crc kubenswrapper[4787]: I1001 09:53:04.277820 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-699b4d5c4-jdg79" event={"ID":"484a5c7d-7633-4e34-a235-af316bbccada","Type":"ContainerStarted","Data":"4e5025b17bff32fc628c55fb38ff4519d21ab89d11e5388d8c4a6cb4392ab891"} Oct 01 09:53:04 crc kubenswrapper[4787]: I1001 09:53:04.279322 4787 generic.go:334] "Generic (PLEG): container finished" podID="17fa3c34-c304-4972-b023-879f3ee28416" containerID="f931d2dda239ceee3044a948206acf3182370a53555e71768aa5a54e24c15e01" exitCode=0 Oct 01 09:53:04 crc kubenswrapper[4787]: I1001 09:53:04.279369 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f4d8965b5-7dw2j" event={"ID":"17fa3c34-c304-4972-b023-879f3ee28416","Type":"ContainerDied","Data":"f931d2dda239ceee3044a948206acf3182370a53555e71768aa5a54e24c15e01"} Oct 01 09:53:04 crc kubenswrapper[4787]: I1001 09:53:04.279389 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f4d8965b5-7dw2j" event={"ID":"17fa3c34-c304-4972-b023-879f3ee28416","Type":"ContainerStarted","Data":"a12d02ab702e8853288c4fdf23cb0e7cbf037021907fee6fea74d55f3f557d66"} Oct 01 09:53:04 crc kubenswrapper[4787]: I1001 09:53:04.287401 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.699407217 podStartE2EDuration="5.287387257s" podCreationTimestamp="2025-10-01 09:52:59 +0000 UTC" firstStartedPulling="2025-10-01 09:53:00.189562798 +0000 UTC m=+1012.304706955" lastFinishedPulling="2025-10-01 09:53:03.777542838 +0000 UTC m=+1015.892686995" observedRunningTime="2025-10-01 09:53:04.285865788 +0000 UTC m=+1016.401009955" watchObservedRunningTime="2025-10-01 09:53:04.287387257 +0000 UTC m=+1016.402531414" Oct 01 09:53:04 crc kubenswrapper[4787]: I1001 09:53:04.294628 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c4db58965-h4fwp" Oct 01 09:53:04 crc kubenswrapper[4787]: I1001 09:53:04.295031 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-79b69df69b-mfcjt" event={"ID":"0244e33a-702c-412d-8074-a2c910f375d0","Type":"ContainerStarted","Data":"f6e497aeeaf590b497d72aedd95d80d7aecc82ff93802c4e050cb6ed19e5e13c"} Oct 01 09:53:04 crc kubenswrapper[4787]: I1001 09:53:04.295136 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-79b69df69b-mfcjt" event={"ID":"0244e33a-702c-412d-8074-a2c910f375d0","Type":"ContainerStarted","Data":"5aed0b8e30923bdfcedfb7b6f9d63cadbeed4c395541d46dd2ec706babcdfe89"} Oct 01 09:53:04 crc kubenswrapper[4787]: I1001 09:53:04.479401 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-c4db58965-h4fwp"] Oct 01 09:53:04 crc kubenswrapper[4787]: I1001 09:53:04.489161 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-c4db58965-h4fwp"] Oct 01 09:53:04 crc kubenswrapper[4787]: I1001 09:53:04.536328 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7835443-3da0-4c3f-8643-b07cbf790a17" path="/var/lib/kubelet/pods/a7835443-3da0-4c3f-8643-b07cbf790a17/volumes" Oct 01 09:53:05 crc kubenswrapper[4787]: I1001 09:53:05.306295 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f4d8965b5-7dw2j" event={"ID":"17fa3c34-c304-4972-b023-879f3ee28416","Type":"ContainerStarted","Data":"1f31d9c2da9765e7b68759a10bde045a5fa9ebda788415b3850019ea6831e3eb"} Oct 01 09:53:05 crc kubenswrapper[4787]: I1001 09:53:05.306603 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-f4d8965b5-7dw2j" Oct 01 09:53:05 crc kubenswrapper[4787]: I1001 09:53:05.315975 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-79b69df69b-mfcjt" event={"ID":"0244e33a-702c-412d-8074-a2c910f375d0","Type":"ContainerStarted","Data":"9809e8583584726a8893b9ddf2134ade7d5b049f4935bc46a74fc053a250125d"} Oct 01 09:53:05 crc kubenswrapper[4787]: I1001 09:53:05.316016 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-79b69df69b-mfcjt" Oct 01 09:53:05 crc kubenswrapper[4787]: I1001 09:53:05.316043 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-79b69df69b-mfcjt" Oct 01 09:53:05 crc kubenswrapper[4787]: I1001 09:53:05.329794 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-887c78f44-f74pm"] Oct 01 09:53:05 crc kubenswrapper[4787]: E1001 09:53:05.330179 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7835443-3da0-4c3f-8643-b07cbf790a17" containerName="init" Oct 01 09:53:05 crc kubenswrapper[4787]: I1001 09:53:05.330195 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7835443-3da0-4c3f-8643-b07cbf790a17" containerName="init" Oct 01 09:53:05 crc kubenswrapper[4787]: E1001 09:53:05.330209 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7835443-3da0-4c3f-8643-b07cbf790a17" containerName="dnsmasq-dns" Oct 01 09:53:05 crc kubenswrapper[4787]: I1001 09:53:05.330216 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7835443-3da0-4c3f-8643-b07cbf790a17" containerName="dnsmasq-dns" Oct 01 09:53:05 crc kubenswrapper[4787]: I1001 09:53:05.330419 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7835443-3da0-4c3f-8643-b07cbf790a17" containerName="dnsmasq-dns" Oct 01 09:53:05 crc kubenswrapper[4787]: I1001 09:53:05.331346 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-887c78f44-f74pm" Oct 01 09:53:05 crc kubenswrapper[4787]: I1001 09:53:05.338556 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 01 09:53:05 crc kubenswrapper[4787]: I1001 09:53:05.338800 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 01 09:53:05 crc kubenswrapper[4787]: I1001 09:53:05.339964 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-f4d8965b5-7dw2j" podStartSLOduration=3.339951319 podStartE2EDuration="3.339951319s" podCreationTimestamp="2025-10-01 09:53:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:53:05.338982944 +0000 UTC m=+1017.454127111" watchObservedRunningTime="2025-10-01 09:53:05.339951319 +0000 UTC m=+1017.455095476" Oct 01 09:53:05 crc kubenswrapper[4787]: I1001 09:53:05.354244 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-887c78f44-f74pm"] Oct 01 09:53:05 crc kubenswrapper[4787]: I1001 09:53:05.467841 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f52bm\" (UniqueName: \"kubernetes.io/projected/873541c7-e7e7-4219-8ce0-c8d5edfbe308-kube-api-access-f52bm\") pod \"barbican-api-887c78f44-f74pm\" (UID: \"873541c7-e7e7-4219-8ce0-c8d5edfbe308\") " pod="openstack/barbican-api-887c78f44-f74pm" Oct 01 09:53:05 crc kubenswrapper[4787]: I1001 09:53:05.467985 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/873541c7-e7e7-4219-8ce0-c8d5edfbe308-internal-tls-certs\") pod \"barbican-api-887c78f44-f74pm\" (UID: \"873541c7-e7e7-4219-8ce0-c8d5edfbe308\") " pod="openstack/barbican-api-887c78f44-f74pm" Oct 01 09:53:05 crc kubenswrapper[4787]: I1001 09:53:05.468024 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/873541c7-e7e7-4219-8ce0-c8d5edfbe308-config-data\") pod \"barbican-api-887c78f44-f74pm\" (UID: \"873541c7-e7e7-4219-8ce0-c8d5edfbe308\") " pod="openstack/barbican-api-887c78f44-f74pm" Oct 01 09:53:05 crc kubenswrapper[4787]: I1001 09:53:05.468066 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/873541c7-e7e7-4219-8ce0-c8d5edfbe308-public-tls-certs\") pod \"barbican-api-887c78f44-f74pm\" (UID: \"873541c7-e7e7-4219-8ce0-c8d5edfbe308\") " pod="openstack/barbican-api-887c78f44-f74pm" Oct 01 09:53:05 crc kubenswrapper[4787]: I1001 09:53:05.468110 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/873541c7-e7e7-4219-8ce0-c8d5edfbe308-logs\") pod \"barbican-api-887c78f44-f74pm\" (UID: \"873541c7-e7e7-4219-8ce0-c8d5edfbe308\") " pod="openstack/barbican-api-887c78f44-f74pm" Oct 01 09:53:05 crc kubenswrapper[4787]: I1001 09:53:05.468144 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/873541c7-e7e7-4219-8ce0-c8d5edfbe308-config-data-custom\") pod \"barbican-api-887c78f44-f74pm\" (UID: \"873541c7-e7e7-4219-8ce0-c8d5edfbe308\") " pod="openstack/barbican-api-887c78f44-f74pm" Oct 01 09:53:05 crc kubenswrapper[4787]: I1001 09:53:05.468170 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/873541c7-e7e7-4219-8ce0-c8d5edfbe308-combined-ca-bundle\") pod \"barbican-api-887c78f44-f74pm\" (UID: \"873541c7-e7e7-4219-8ce0-c8d5edfbe308\") " pod="openstack/barbican-api-887c78f44-f74pm" Oct 01 09:53:05 crc kubenswrapper[4787]: I1001 09:53:05.572259 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/873541c7-e7e7-4219-8ce0-c8d5edfbe308-logs\") pod \"barbican-api-887c78f44-f74pm\" (UID: \"873541c7-e7e7-4219-8ce0-c8d5edfbe308\") " pod="openstack/barbican-api-887c78f44-f74pm" Oct 01 09:53:05 crc kubenswrapper[4787]: I1001 09:53:05.572320 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/873541c7-e7e7-4219-8ce0-c8d5edfbe308-config-data-custom\") pod \"barbican-api-887c78f44-f74pm\" (UID: \"873541c7-e7e7-4219-8ce0-c8d5edfbe308\") " pod="openstack/barbican-api-887c78f44-f74pm" Oct 01 09:53:05 crc kubenswrapper[4787]: I1001 09:53:05.572360 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/873541c7-e7e7-4219-8ce0-c8d5edfbe308-combined-ca-bundle\") pod \"barbican-api-887c78f44-f74pm\" (UID: \"873541c7-e7e7-4219-8ce0-c8d5edfbe308\") " pod="openstack/barbican-api-887c78f44-f74pm" Oct 01 09:53:05 crc kubenswrapper[4787]: I1001 09:53:05.572432 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f52bm\" (UniqueName: \"kubernetes.io/projected/873541c7-e7e7-4219-8ce0-c8d5edfbe308-kube-api-access-f52bm\") pod \"barbican-api-887c78f44-f74pm\" (UID: \"873541c7-e7e7-4219-8ce0-c8d5edfbe308\") " pod="openstack/barbican-api-887c78f44-f74pm" Oct 01 09:53:05 crc kubenswrapper[4787]: I1001 09:53:05.572500 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/873541c7-e7e7-4219-8ce0-c8d5edfbe308-internal-tls-certs\") pod \"barbican-api-887c78f44-f74pm\" (UID: \"873541c7-e7e7-4219-8ce0-c8d5edfbe308\") " pod="openstack/barbican-api-887c78f44-f74pm" Oct 01 09:53:05 crc kubenswrapper[4787]: I1001 09:53:05.572535 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/873541c7-e7e7-4219-8ce0-c8d5edfbe308-config-data\") pod \"barbican-api-887c78f44-f74pm\" (UID: \"873541c7-e7e7-4219-8ce0-c8d5edfbe308\") " pod="openstack/barbican-api-887c78f44-f74pm" Oct 01 09:53:05 crc kubenswrapper[4787]: I1001 09:53:05.572572 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/873541c7-e7e7-4219-8ce0-c8d5edfbe308-public-tls-certs\") pod \"barbican-api-887c78f44-f74pm\" (UID: \"873541c7-e7e7-4219-8ce0-c8d5edfbe308\") " pod="openstack/barbican-api-887c78f44-f74pm" Oct 01 09:53:05 crc kubenswrapper[4787]: I1001 09:53:05.573783 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/873541c7-e7e7-4219-8ce0-c8d5edfbe308-logs\") pod \"barbican-api-887c78f44-f74pm\" (UID: \"873541c7-e7e7-4219-8ce0-c8d5edfbe308\") " pod="openstack/barbican-api-887c78f44-f74pm" Oct 01 09:53:05 crc kubenswrapper[4787]: I1001 09:53:05.582883 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/873541c7-e7e7-4219-8ce0-c8d5edfbe308-public-tls-certs\") pod \"barbican-api-887c78f44-f74pm\" (UID: \"873541c7-e7e7-4219-8ce0-c8d5edfbe308\") " pod="openstack/barbican-api-887c78f44-f74pm" Oct 01 09:53:05 crc kubenswrapper[4787]: I1001 09:53:05.583175 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/873541c7-e7e7-4219-8ce0-c8d5edfbe308-config-data\") pod \"barbican-api-887c78f44-f74pm\" (UID: \"873541c7-e7e7-4219-8ce0-c8d5edfbe308\") " pod="openstack/barbican-api-887c78f44-f74pm" Oct 01 09:53:05 crc kubenswrapper[4787]: I1001 09:53:05.585563 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/873541c7-e7e7-4219-8ce0-c8d5edfbe308-internal-tls-certs\") pod \"barbican-api-887c78f44-f74pm\" (UID: \"873541c7-e7e7-4219-8ce0-c8d5edfbe308\") " pod="openstack/barbican-api-887c78f44-f74pm" Oct 01 09:53:05 crc kubenswrapper[4787]: I1001 09:53:05.585927 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/873541c7-e7e7-4219-8ce0-c8d5edfbe308-combined-ca-bundle\") pod \"barbican-api-887c78f44-f74pm\" (UID: \"873541c7-e7e7-4219-8ce0-c8d5edfbe308\") " pod="openstack/barbican-api-887c78f44-f74pm" Oct 01 09:53:05 crc kubenswrapper[4787]: I1001 09:53:05.603838 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/873541c7-e7e7-4219-8ce0-c8d5edfbe308-config-data-custom\") pod \"barbican-api-887c78f44-f74pm\" (UID: \"873541c7-e7e7-4219-8ce0-c8d5edfbe308\") " pod="openstack/barbican-api-887c78f44-f74pm" Oct 01 09:53:05 crc kubenswrapper[4787]: I1001 09:53:05.623126 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f52bm\" (UniqueName: \"kubernetes.io/projected/873541c7-e7e7-4219-8ce0-c8d5edfbe308-kube-api-access-f52bm\") pod \"barbican-api-887c78f44-f74pm\" (UID: \"873541c7-e7e7-4219-8ce0-c8d5edfbe308\") " pod="openstack/barbican-api-887c78f44-f74pm" Oct 01 09:53:05 crc kubenswrapper[4787]: I1001 09:53:05.672829 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-887c78f44-f74pm" Oct 01 09:53:06 crc kubenswrapper[4787]: I1001 09:53:06.334599 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-7f57487985-p72bt" Oct 01 09:53:06 crc kubenswrapper[4787]: I1001 09:53:06.364621 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-79b69df69b-mfcjt" podStartSLOduration=4.364598674 podStartE2EDuration="4.364598674s" podCreationTimestamp="2025-10-01 09:53:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:53:05.423520605 +0000 UTC m=+1017.538664772" watchObservedRunningTime="2025-10-01 09:53:06.364598674 +0000 UTC m=+1018.479742831" Oct 01 09:53:06 crc kubenswrapper[4787]: I1001 09:53:06.672757 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-887c78f44-f74pm"] Oct 01 09:53:07 crc kubenswrapper[4787]: I1001 09:53:07.341111 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-699b4d5c4-jdg79" event={"ID":"484a5c7d-7633-4e34-a235-af316bbccada","Type":"ContainerStarted","Data":"aa0c68f7d2459cc3d8ce3f7c73f11f3aabfd3f7eb8d0dce3fc791bb2d20876fd"} Oct 01 09:53:07 crc kubenswrapper[4787]: I1001 09:53:07.341509 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-699b4d5c4-jdg79" event={"ID":"484a5c7d-7633-4e34-a235-af316bbccada","Type":"ContainerStarted","Data":"0b264ac84c48278affe3b7c2dd23f7a5cbf60683ec6e8fae010a7db0e3578315"} Oct 01 09:53:07 crc kubenswrapper[4787]: I1001 09:53:07.344704 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-887c78f44-f74pm" event={"ID":"873541c7-e7e7-4219-8ce0-c8d5edfbe308","Type":"ContainerStarted","Data":"0368e418540fbcf8086c6abe897d919607205db831d3dd89c3ea966dd7f49ef5"} Oct 01 09:53:07 crc kubenswrapper[4787]: I1001 09:53:07.344747 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-887c78f44-f74pm" event={"ID":"873541c7-e7e7-4219-8ce0-c8d5edfbe308","Type":"ContainerStarted","Data":"9b098fcbf24648a65867171d6ed5d910acb20bd03bd2d882fce99773445e71d4"} Oct 01 09:53:07 crc kubenswrapper[4787]: I1001 09:53:07.344757 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-887c78f44-f74pm" event={"ID":"873541c7-e7e7-4219-8ce0-c8d5edfbe308","Type":"ContainerStarted","Data":"272fd5536dcc8813623fe33200c8dbea175691569579746269e08f775f0b3430"} Oct 01 09:53:07 crc kubenswrapper[4787]: I1001 09:53:07.345430 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-887c78f44-f74pm" Oct 01 09:53:07 crc kubenswrapper[4787]: I1001 09:53:07.345653 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-887c78f44-f74pm" Oct 01 09:53:07 crc kubenswrapper[4787]: I1001 09:53:07.348296 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-64d9c8b8f9-zv7w7" event={"ID":"ea69f2cb-9d4f-4b3c-a0ba-8a3fc8fce274","Type":"ContainerStarted","Data":"afe49d32ebcdc36eb25b6f3d76b45d8f8859c16fc020a0dfbeadecaca8eaa431"} Oct 01 09:53:07 crc kubenswrapper[4787]: I1001 09:53:07.348439 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-64d9c8b8f9-zv7w7" event={"ID":"ea69f2cb-9d4f-4b3c-a0ba-8a3fc8fce274","Type":"ContainerStarted","Data":"cf835cbd15dd3625471f0dc074141cb63fb8aec0528253886ef7abe704b47bd7"} Oct 01 09:53:07 crc kubenswrapper[4787]: I1001 09:53:07.407610 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-699b4d5c4-jdg79" podStartSLOduration=2.726782523 podStartE2EDuration="5.407567212s" podCreationTimestamp="2025-10-01 09:53:02 +0000 UTC" firstStartedPulling="2025-10-01 09:53:03.505385467 +0000 UTC m=+1015.620529624" lastFinishedPulling="2025-10-01 09:53:06.186170156 +0000 UTC m=+1018.301314313" observedRunningTime="2025-10-01 09:53:07.376433764 +0000 UTC m=+1019.491577941" watchObservedRunningTime="2025-10-01 09:53:07.407567212 +0000 UTC m=+1019.522711369" Oct 01 09:53:07 crc kubenswrapper[4787]: I1001 09:53:07.418113 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-887c78f44-f74pm" podStartSLOduration=2.418090268 podStartE2EDuration="2.418090268s" podCreationTimestamp="2025-10-01 09:53:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:53:07.399933269 +0000 UTC m=+1019.515077426" watchObservedRunningTime="2025-10-01 09:53:07.418090268 +0000 UTC m=+1019.533234425" Oct 01 09:53:07 crc kubenswrapper[4787]: I1001 09:53:07.434005 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-64d9c8b8f9-zv7w7" podStartSLOduration=2.7107066570000002 podStartE2EDuration="5.433987482s" podCreationTimestamp="2025-10-01 09:53:02 +0000 UTC" firstStartedPulling="2025-10-01 09:53:03.468115513 +0000 UTC m=+1015.583259670" lastFinishedPulling="2025-10-01 09:53:06.191396338 +0000 UTC m=+1018.306540495" observedRunningTime="2025-10-01 09:53:07.432121014 +0000 UTC m=+1019.547265191" watchObservedRunningTime="2025-10-01 09:53:07.433987482 +0000 UTC m=+1019.549131639" Oct 01 09:53:07 crc kubenswrapper[4787]: I1001 09:53:07.515803 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 01 09:53:07 crc kubenswrapper[4787]: I1001 09:53:07.517219 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 01 09:53:07 crc kubenswrapper[4787]: I1001 09:53:07.519779 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 01 09:53:07 crc kubenswrapper[4787]: I1001 09:53:07.519974 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 01 09:53:07 crc kubenswrapper[4787]: I1001 09:53:07.520537 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-lxmhp" Oct 01 09:53:07 crc kubenswrapper[4787]: I1001 09:53:07.528571 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 01 09:53:07 crc kubenswrapper[4787]: I1001 09:53:07.623018 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9af8596f-452d-4824-90a2-9eefc13f2428-combined-ca-bundle\") pod \"openstackclient\" (UID: \"9af8596f-452d-4824-90a2-9eefc13f2428\") " pod="openstack/openstackclient" Oct 01 09:53:07 crc kubenswrapper[4787]: I1001 09:53:07.623090 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfq99\" (UniqueName: \"kubernetes.io/projected/9af8596f-452d-4824-90a2-9eefc13f2428-kube-api-access-kfq99\") pod \"openstackclient\" (UID: \"9af8596f-452d-4824-90a2-9eefc13f2428\") " pod="openstack/openstackclient" Oct 01 09:53:07 crc kubenswrapper[4787]: I1001 09:53:07.623161 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9af8596f-452d-4824-90a2-9eefc13f2428-openstack-config\") pod \"openstackclient\" (UID: \"9af8596f-452d-4824-90a2-9eefc13f2428\") " pod="openstack/openstackclient" Oct 01 09:53:07 crc kubenswrapper[4787]: I1001 09:53:07.623184 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9af8596f-452d-4824-90a2-9eefc13f2428-openstack-config-secret\") pod \"openstackclient\" (UID: \"9af8596f-452d-4824-90a2-9eefc13f2428\") " pod="openstack/openstackclient" Oct 01 09:53:07 crc kubenswrapper[4787]: I1001 09:53:07.713293 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 01 09:53:07 crc kubenswrapper[4787]: E1001 09:53:07.714062 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle kube-api-access-kfq99 openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/openstackclient" podUID="9af8596f-452d-4824-90a2-9eefc13f2428" Oct 01 09:53:07 crc kubenswrapper[4787]: I1001 09:53:07.724773 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 01 09:53:07 crc kubenswrapper[4787]: I1001 09:53:07.724952 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9af8596f-452d-4824-90a2-9eefc13f2428-openstack-config-secret\") pod \"openstackclient\" (UID: \"9af8596f-452d-4824-90a2-9eefc13f2428\") " pod="openstack/openstackclient" Oct 01 09:53:07 crc kubenswrapper[4787]: I1001 09:53:07.725122 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9af8596f-452d-4824-90a2-9eefc13f2428-combined-ca-bundle\") pod \"openstackclient\" (UID: \"9af8596f-452d-4824-90a2-9eefc13f2428\") " pod="openstack/openstackclient" Oct 01 09:53:07 crc kubenswrapper[4787]: I1001 09:53:07.725145 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfq99\" (UniqueName: \"kubernetes.io/projected/9af8596f-452d-4824-90a2-9eefc13f2428-kube-api-access-kfq99\") pod \"openstackclient\" (UID: \"9af8596f-452d-4824-90a2-9eefc13f2428\") " pod="openstack/openstackclient" Oct 01 09:53:07 crc kubenswrapper[4787]: I1001 09:53:07.725201 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9af8596f-452d-4824-90a2-9eefc13f2428-openstack-config\") pod \"openstackclient\" (UID: \"9af8596f-452d-4824-90a2-9eefc13f2428\") " pod="openstack/openstackclient" Oct 01 09:53:07 crc kubenswrapper[4787]: I1001 09:53:07.726131 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9af8596f-452d-4824-90a2-9eefc13f2428-openstack-config\") pod \"openstackclient\" (UID: \"9af8596f-452d-4824-90a2-9eefc13f2428\") " pod="openstack/openstackclient" Oct 01 09:53:07 crc kubenswrapper[4787]: E1001 09:53:07.729710 4787 projected.go:194] Error preparing data for projected volume kube-api-access-kfq99 for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: User "system:node:crc" cannot create resource "serviceaccounts/token" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Oct 01 09:53:07 crc kubenswrapper[4787]: E1001 09:53:07.729765 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9af8596f-452d-4824-90a2-9eefc13f2428-kube-api-access-kfq99 podName:9af8596f-452d-4824-90a2-9eefc13f2428 nodeName:}" failed. No retries permitted until 2025-10-01 09:53:08.22975027 +0000 UTC m=+1020.344894427 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-kfq99" (UniqueName: "kubernetes.io/projected/9af8596f-452d-4824-90a2-9eefc13f2428-kube-api-access-kfq99") pod "openstackclient" (UID: "9af8596f-452d-4824-90a2-9eefc13f2428") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: User "system:node:crc" cannot create resource "serviceaccounts/token" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Oct 01 09:53:07 crc kubenswrapper[4787]: I1001 09:53:07.733784 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9af8596f-452d-4824-90a2-9eefc13f2428-combined-ca-bundle\") pod \"openstackclient\" (UID: \"9af8596f-452d-4824-90a2-9eefc13f2428\") " pod="openstack/openstackclient" Oct 01 09:53:07 crc kubenswrapper[4787]: I1001 09:53:07.734137 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9af8596f-452d-4824-90a2-9eefc13f2428-openstack-config-secret\") pod \"openstackclient\" (UID: \"9af8596f-452d-4824-90a2-9eefc13f2428\") " pod="openstack/openstackclient" Oct 01 09:53:07 crc kubenswrapper[4787]: I1001 09:53:07.786910 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 01 09:53:07 crc kubenswrapper[4787]: I1001 09:53:07.789514 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 01 09:53:07 crc kubenswrapper[4787]: I1001 09:53:07.835624 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 01 09:53:07 crc kubenswrapper[4787]: I1001 09:53:07.928495 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ccb620ed-ad28-4ec7-8319-6d5bb3ce10ec-openstack-config\") pod \"openstackclient\" (UID: \"ccb620ed-ad28-4ec7-8319-6d5bb3ce10ec\") " pod="openstack/openstackclient" Oct 01 09:53:07 crc kubenswrapper[4787]: I1001 09:53:07.928598 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccb620ed-ad28-4ec7-8319-6d5bb3ce10ec-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ccb620ed-ad28-4ec7-8319-6d5bb3ce10ec\") " pod="openstack/openstackclient" Oct 01 09:53:07 crc kubenswrapper[4787]: I1001 09:53:07.928617 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9rhg\" (UniqueName: \"kubernetes.io/projected/ccb620ed-ad28-4ec7-8319-6d5bb3ce10ec-kube-api-access-v9rhg\") pod \"openstackclient\" (UID: \"ccb620ed-ad28-4ec7-8319-6d5bb3ce10ec\") " pod="openstack/openstackclient" Oct 01 09:53:07 crc kubenswrapper[4787]: I1001 09:53:07.928676 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ccb620ed-ad28-4ec7-8319-6d5bb3ce10ec-openstack-config-secret\") pod \"openstackclient\" (UID: \"ccb620ed-ad28-4ec7-8319-6d5bb3ce10ec\") " pod="openstack/openstackclient" Oct 01 09:53:08 crc kubenswrapper[4787]: I1001 09:53:08.030035 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ccb620ed-ad28-4ec7-8319-6d5bb3ce10ec-openstack-config\") pod \"openstackclient\" (UID: \"ccb620ed-ad28-4ec7-8319-6d5bb3ce10ec\") " pod="openstack/openstackclient" Oct 01 09:53:08 crc kubenswrapper[4787]: I1001 09:53:08.030144 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccb620ed-ad28-4ec7-8319-6d5bb3ce10ec-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ccb620ed-ad28-4ec7-8319-6d5bb3ce10ec\") " pod="openstack/openstackclient" Oct 01 09:53:08 crc kubenswrapper[4787]: I1001 09:53:08.030200 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9rhg\" (UniqueName: \"kubernetes.io/projected/ccb620ed-ad28-4ec7-8319-6d5bb3ce10ec-kube-api-access-v9rhg\") pod \"openstackclient\" (UID: \"ccb620ed-ad28-4ec7-8319-6d5bb3ce10ec\") " pod="openstack/openstackclient" Oct 01 09:53:08 crc kubenswrapper[4787]: I1001 09:53:08.030657 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ccb620ed-ad28-4ec7-8319-6d5bb3ce10ec-openstack-config-secret\") pod \"openstackclient\" (UID: \"ccb620ed-ad28-4ec7-8319-6d5bb3ce10ec\") " pod="openstack/openstackclient" Oct 01 09:53:08 crc kubenswrapper[4787]: I1001 09:53:08.031017 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ccb620ed-ad28-4ec7-8319-6d5bb3ce10ec-openstack-config\") pod \"openstackclient\" (UID: \"ccb620ed-ad28-4ec7-8319-6d5bb3ce10ec\") " pod="openstack/openstackclient" Oct 01 09:53:08 crc kubenswrapper[4787]: I1001 09:53:08.035602 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ccb620ed-ad28-4ec7-8319-6d5bb3ce10ec-openstack-config-secret\") pod \"openstackclient\" (UID: \"ccb620ed-ad28-4ec7-8319-6d5bb3ce10ec\") " pod="openstack/openstackclient" Oct 01 09:53:08 crc kubenswrapper[4787]: I1001 09:53:08.035948 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccb620ed-ad28-4ec7-8319-6d5bb3ce10ec-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ccb620ed-ad28-4ec7-8319-6d5bb3ce10ec\") " pod="openstack/openstackclient" Oct 01 09:53:08 crc kubenswrapper[4787]: I1001 09:53:08.071423 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9rhg\" (UniqueName: \"kubernetes.io/projected/ccb620ed-ad28-4ec7-8319-6d5bb3ce10ec-kube-api-access-v9rhg\") pod \"openstackclient\" (UID: \"ccb620ed-ad28-4ec7-8319-6d5bb3ce10ec\") " pod="openstack/openstackclient" Oct 01 09:53:08 crc kubenswrapper[4787]: I1001 09:53:08.159017 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 01 09:53:08 crc kubenswrapper[4787]: I1001 09:53:08.234990 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfq99\" (UniqueName: \"kubernetes.io/projected/9af8596f-452d-4824-90a2-9eefc13f2428-kube-api-access-kfq99\") pod \"openstackclient\" (UID: \"9af8596f-452d-4824-90a2-9eefc13f2428\") " pod="openstack/openstackclient" Oct 01 09:53:08 crc kubenswrapper[4787]: E1001 09:53:08.237781 4787 projected.go:194] Error preparing data for projected volume kube-api-access-kfq99 for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (9af8596f-452d-4824-90a2-9eefc13f2428) does not match the UID in record. The object might have been deleted and then recreated Oct 01 09:53:08 crc kubenswrapper[4787]: E1001 09:53:08.237913 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9af8596f-452d-4824-90a2-9eefc13f2428-kube-api-access-kfq99 podName:9af8596f-452d-4824-90a2-9eefc13f2428 nodeName:}" failed. No retries permitted until 2025-10-01 09:53:09.237877637 +0000 UTC m=+1021.353021804 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-kfq99" (UniqueName: "kubernetes.io/projected/9af8596f-452d-4824-90a2-9eefc13f2428-kube-api-access-kfq99") pod "openstackclient" (UID: "9af8596f-452d-4824-90a2-9eefc13f2428") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (9af8596f-452d-4824-90a2-9eefc13f2428) does not match the UID in record. The object might have been deleted and then recreated Oct 01 09:53:08 crc kubenswrapper[4787]: I1001 09:53:08.366166 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 01 09:53:08 crc kubenswrapper[4787]: I1001 09:53:08.370182 4787 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="9af8596f-452d-4824-90a2-9eefc13f2428" podUID="ccb620ed-ad28-4ec7-8319-6d5bb3ce10ec" Oct 01 09:53:08 crc kubenswrapper[4787]: I1001 09:53:08.382382 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 01 09:53:08 crc kubenswrapper[4787]: I1001 09:53:08.550776 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9af8596f-452d-4824-90a2-9eefc13f2428-openstack-config-secret\") pod \"9af8596f-452d-4824-90a2-9eefc13f2428\" (UID: \"9af8596f-452d-4824-90a2-9eefc13f2428\") " Oct 01 09:53:08 crc kubenswrapper[4787]: I1001 09:53:08.551204 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9af8596f-452d-4824-90a2-9eefc13f2428-openstack-config\") pod \"9af8596f-452d-4824-90a2-9eefc13f2428\" (UID: \"9af8596f-452d-4824-90a2-9eefc13f2428\") " Oct 01 09:53:08 crc kubenswrapper[4787]: I1001 09:53:08.551282 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9af8596f-452d-4824-90a2-9eefc13f2428-combined-ca-bundle\") pod \"9af8596f-452d-4824-90a2-9eefc13f2428\" (UID: \"9af8596f-452d-4824-90a2-9eefc13f2428\") " Oct 01 09:53:08 crc kubenswrapper[4787]: I1001 09:53:08.551716 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfq99\" (UniqueName: \"kubernetes.io/projected/9af8596f-452d-4824-90a2-9eefc13f2428-kube-api-access-kfq99\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:08 crc kubenswrapper[4787]: I1001 09:53:08.555447 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9af8596f-452d-4824-90a2-9eefc13f2428-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "9af8596f-452d-4824-90a2-9eefc13f2428" (UID: "9af8596f-452d-4824-90a2-9eefc13f2428"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:53:08 crc kubenswrapper[4787]: I1001 09:53:08.577061 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9af8596f-452d-4824-90a2-9eefc13f2428-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "9af8596f-452d-4824-90a2-9eefc13f2428" (UID: "9af8596f-452d-4824-90a2-9eefc13f2428"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:08 crc kubenswrapper[4787]: I1001 09:53:08.585199 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9af8596f-452d-4824-90a2-9eefc13f2428-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9af8596f-452d-4824-90a2-9eefc13f2428" (UID: "9af8596f-452d-4824-90a2-9eefc13f2428"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:08 crc kubenswrapper[4787]: I1001 09:53:08.653202 4787 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9af8596f-452d-4824-90a2-9eefc13f2428-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:08 crc kubenswrapper[4787]: I1001 09:53:08.653300 4787 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9af8596f-452d-4824-90a2-9eefc13f2428-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:08 crc kubenswrapper[4787]: I1001 09:53:08.653310 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9af8596f-452d-4824-90a2-9eefc13f2428-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:08 crc kubenswrapper[4787]: I1001 09:53:08.738022 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 01 09:53:08 crc kubenswrapper[4787]: W1001 09:53:08.744271 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podccb620ed_ad28_4ec7_8319_6d5bb3ce10ec.slice/crio-68e00316912f310be8e87a0e74103c60d48efbfcd44688b4f340881fb52c9541 WatchSource:0}: Error finding container 68e00316912f310be8e87a0e74103c60d48efbfcd44688b4f340881fb52c9541: Status 404 returned error can't find the container with id 68e00316912f310be8e87a0e74103c60d48efbfcd44688b4f340881fb52c9541 Oct 01 09:53:09 crc kubenswrapper[4787]: I1001 09:53:09.373700 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 01 09:53:09 crc kubenswrapper[4787]: I1001 09:53:09.373891 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"ccb620ed-ad28-4ec7-8319-6d5bb3ce10ec","Type":"ContainerStarted","Data":"68e00316912f310be8e87a0e74103c60d48efbfcd44688b4f340881fb52c9541"} Oct 01 09:53:09 crc kubenswrapper[4787]: I1001 09:53:09.390753 4787 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="9af8596f-452d-4824-90a2-9eefc13f2428" podUID="ccb620ed-ad28-4ec7-8319-6d5bb3ce10ec" Oct 01 09:53:10 crc kubenswrapper[4787]: I1001 09:53:10.533731 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9af8596f-452d-4824-90a2-9eefc13f2428" path="/var/lib/kubelet/pods/9af8596f-452d-4824-90a2-9eefc13f2428/volumes" Oct 01 09:53:11 crc kubenswrapper[4787]: I1001 09:53:11.396909 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-8mrcw" event={"ID":"ca8b9373-e5e5-4dc1-8995-44ae47abbb13","Type":"ContainerStarted","Data":"636e18451477b045320e69118c705fe6303e8717966fab13f8b4ae36f44898f5"} Oct 01 09:53:11 crc kubenswrapper[4787]: I1001 09:53:11.421815 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-8mrcw" podStartSLOduration=3.355604232 podStartE2EDuration="43.421794115s" podCreationTimestamp="2025-10-01 09:52:28 +0000 UTC" firstStartedPulling="2025-10-01 09:52:29.933012592 +0000 UTC m=+982.048156749" lastFinishedPulling="2025-10-01 09:53:09.999202475 +0000 UTC m=+1022.114346632" observedRunningTime="2025-10-01 09:53:11.414145131 +0000 UTC m=+1023.529289298" watchObservedRunningTime="2025-10-01 09:53:11.421794115 +0000 UTC m=+1023.536938282" Oct 01 09:53:12 crc kubenswrapper[4787]: I1001 09:53:12.080242 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:53:12 crc kubenswrapper[4787]: I1001 09:53:12.081458 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c" containerName="ceilometer-central-agent" containerID="cri-o://98c33f6c04008cd30bdef34a1de003ea58ef86bd06cb44398a6e45d9e7916bc5" gracePeriod=30 Oct 01 09:53:12 crc kubenswrapper[4787]: I1001 09:53:12.082098 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c" containerName="sg-core" containerID="cri-o://9ea5042ebd4a4cfdce6dd4dd8b3f130111459a9b369f9c3d2aa1cb0f72d20ad0" gracePeriod=30 Oct 01 09:53:12 crc kubenswrapper[4787]: I1001 09:53:12.082143 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c" containerName="ceilometer-notification-agent" containerID="cri-o://974cfa14c1d8f423c33e89b90de41d56b73acf24136c57d8454283e7000a8425" gracePeriod=30 Oct 01 09:53:12 crc kubenswrapper[4787]: I1001 09:53:12.082357 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c" containerName="proxy-httpd" containerID="cri-o://ae68e6f7fc578d6c34f91b087188337008568de65db874e1e8a7966190fdc84e" gracePeriod=30 Oct 01 09:53:12 crc kubenswrapper[4787]: I1001 09:53:12.408196 4787 generic.go:334] "Generic (PLEG): container finished" podID="b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c" containerID="ae68e6f7fc578d6c34f91b087188337008568de65db874e1e8a7966190fdc84e" exitCode=0 Oct 01 09:53:12 crc kubenswrapper[4787]: I1001 09:53:12.408548 4787 generic.go:334] "Generic (PLEG): container finished" podID="b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c" containerID="9ea5042ebd4a4cfdce6dd4dd8b3f130111459a9b369f9c3d2aa1cb0f72d20ad0" exitCode=2 Oct 01 09:53:12 crc kubenswrapper[4787]: I1001 09:53:12.408578 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c","Type":"ContainerDied","Data":"ae68e6f7fc578d6c34f91b087188337008568de65db874e1e8a7966190fdc84e"} Oct 01 09:53:12 crc kubenswrapper[4787]: I1001 09:53:12.408610 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c","Type":"ContainerDied","Data":"9ea5042ebd4a4cfdce6dd4dd8b3f130111459a9b369f9c3d2aa1cb0f72d20ad0"} Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.041016 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-5c7b59dd4f-spxm4"] Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.044452 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5c7b59dd4f-spxm4" Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.054845 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.055054 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.055236 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.065466 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5c7b59dd4f-spxm4"] Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.127194 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-f4d8965b5-7dw2j" Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.148115 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e743cd5e-e8c6-4fe2-9480-3a30316b8e23-combined-ca-bundle\") pod \"swift-proxy-5c7b59dd4f-spxm4\" (UID: \"e743cd5e-e8c6-4fe2-9480-3a30316b8e23\") " pod="openstack/swift-proxy-5c7b59dd4f-spxm4" Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.148196 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e743cd5e-e8c6-4fe2-9480-3a30316b8e23-internal-tls-certs\") pod \"swift-proxy-5c7b59dd4f-spxm4\" (UID: \"e743cd5e-e8c6-4fe2-9480-3a30316b8e23\") " pod="openstack/swift-proxy-5c7b59dd4f-spxm4" Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.148280 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e743cd5e-e8c6-4fe2-9480-3a30316b8e23-run-httpd\") pod \"swift-proxy-5c7b59dd4f-spxm4\" (UID: \"e743cd5e-e8c6-4fe2-9480-3a30316b8e23\") " pod="openstack/swift-proxy-5c7b59dd4f-spxm4" Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.148500 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e743cd5e-e8c6-4fe2-9480-3a30316b8e23-etc-swift\") pod \"swift-proxy-5c7b59dd4f-spxm4\" (UID: \"e743cd5e-e8c6-4fe2-9480-3a30316b8e23\") " pod="openstack/swift-proxy-5c7b59dd4f-spxm4" Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.148537 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4dng\" (UniqueName: \"kubernetes.io/projected/e743cd5e-e8c6-4fe2-9480-3a30316b8e23-kube-api-access-d4dng\") pod \"swift-proxy-5c7b59dd4f-spxm4\" (UID: \"e743cd5e-e8c6-4fe2-9480-3a30316b8e23\") " pod="openstack/swift-proxy-5c7b59dd4f-spxm4" Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.148609 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e743cd5e-e8c6-4fe2-9480-3a30316b8e23-log-httpd\") pod \"swift-proxy-5c7b59dd4f-spxm4\" (UID: \"e743cd5e-e8c6-4fe2-9480-3a30316b8e23\") " pod="openstack/swift-proxy-5c7b59dd4f-spxm4" Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.148668 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e743cd5e-e8c6-4fe2-9480-3a30316b8e23-config-data\") pod \"swift-proxy-5c7b59dd4f-spxm4\" (UID: \"e743cd5e-e8c6-4fe2-9480-3a30316b8e23\") " pod="openstack/swift-proxy-5c7b59dd4f-spxm4" Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.148739 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e743cd5e-e8c6-4fe2-9480-3a30316b8e23-public-tls-certs\") pod \"swift-proxy-5c7b59dd4f-spxm4\" (UID: \"e743cd5e-e8c6-4fe2-9480-3a30316b8e23\") " pod="openstack/swift-proxy-5c7b59dd4f-spxm4" Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.177221 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-67879968b-d5598" podUID="7983e213-c317-42b2-8944-85bdb78dac8f" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.209549 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b69d7b58f-t9lc6"] Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.209793 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b69d7b58f-t9lc6" podUID="6808314c-0e06-4fa4-9c3e-269e5952fd3e" containerName="dnsmasq-dns" containerID="cri-o://25a024f9e9cad11ef9f8ce3dd6a938d6dba2bec6ef4e26686340150ab47e10f9" gracePeriod=10 Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.252249 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e743cd5e-e8c6-4fe2-9480-3a30316b8e23-log-httpd\") pod \"swift-proxy-5c7b59dd4f-spxm4\" (UID: \"e743cd5e-e8c6-4fe2-9480-3a30316b8e23\") " pod="openstack/swift-proxy-5c7b59dd4f-spxm4" Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.252358 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e743cd5e-e8c6-4fe2-9480-3a30316b8e23-config-data\") pod \"swift-proxy-5c7b59dd4f-spxm4\" (UID: \"e743cd5e-e8c6-4fe2-9480-3a30316b8e23\") " pod="openstack/swift-proxy-5c7b59dd4f-spxm4" Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.252407 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e743cd5e-e8c6-4fe2-9480-3a30316b8e23-public-tls-certs\") pod \"swift-proxy-5c7b59dd4f-spxm4\" (UID: \"e743cd5e-e8c6-4fe2-9480-3a30316b8e23\") " pod="openstack/swift-proxy-5c7b59dd4f-spxm4" Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.252452 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e743cd5e-e8c6-4fe2-9480-3a30316b8e23-combined-ca-bundle\") pod \"swift-proxy-5c7b59dd4f-spxm4\" (UID: \"e743cd5e-e8c6-4fe2-9480-3a30316b8e23\") " pod="openstack/swift-proxy-5c7b59dd4f-spxm4" Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.252523 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e743cd5e-e8c6-4fe2-9480-3a30316b8e23-internal-tls-certs\") pod \"swift-proxy-5c7b59dd4f-spxm4\" (UID: \"e743cd5e-e8c6-4fe2-9480-3a30316b8e23\") " pod="openstack/swift-proxy-5c7b59dd4f-spxm4" Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.252553 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e743cd5e-e8c6-4fe2-9480-3a30316b8e23-run-httpd\") pod \"swift-proxy-5c7b59dd4f-spxm4\" (UID: \"e743cd5e-e8c6-4fe2-9480-3a30316b8e23\") " pod="openstack/swift-proxy-5c7b59dd4f-spxm4" Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.252678 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e743cd5e-e8c6-4fe2-9480-3a30316b8e23-etc-swift\") pod \"swift-proxy-5c7b59dd4f-spxm4\" (UID: \"e743cd5e-e8c6-4fe2-9480-3a30316b8e23\") " pod="openstack/swift-proxy-5c7b59dd4f-spxm4" Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.252717 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4dng\" (UniqueName: \"kubernetes.io/projected/e743cd5e-e8c6-4fe2-9480-3a30316b8e23-kube-api-access-d4dng\") pod \"swift-proxy-5c7b59dd4f-spxm4\" (UID: \"e743cd5e-e8c6-4fe2-9480-3a30316b8e23\") " pod="openstack/swift-proxy-5c7b59dd4f-spxm4" Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.253577 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e743cd5e-e8c6-4fe2-9480-3a30316b8e23-log-httpd\") pod \"swift-proxy-5c7b59dd4f-spxm4\" (UID: \"e743cd5e-e8c6-4fe2-9480-3a30316b8e23\") " pod="openstack/swift-proxy-5c7b59dd4f-spxm4" Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.256829 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e743cd5e-e8c6-4fe2-9480-3a30316b8e23-run-httpd\") pod \"swift-proxy-5c7b59dd4f-spxm4\" (UID: \"e743cd5e-e8c6-4fe2-9480-3a30316b8e23\") " pod="openstack/swift-proxy-5c7b59dd4f-spxm4" Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.265040 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e743cd5e-e8c6-4fe2-9480-3a30316b8e23-internal-tls-certs\") pod \"swift-proxy-5c7b59dd4f-spxm4\" (UID: \"e743cd5e-e8c6-4fe2-9480-3a30316b8e23\") " pod="openstack/swift-proxy-5c7b59dd4f-spxm4" Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.265173 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e743cd5e-e8c6-4fe2-9480-3a30316b8e23-combined-ca-bundle\") pod \"swift-proxy-5c7b59dd4f-spxm4\" (UID: \"e743cd5e-e8c6-4fe2-9480-3a30316b8e23\") " pod="openstack/swift-proxy-5c7b59dd4f-spxm4" Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.268502 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e743cd5e-e8c6-4fe2-9480-3a30316b8e23-config-data\") pod \"swift-proxy-5c7b59dd4f-spxm4\" (UID: \"e743cd5e-e8c6-4fe2-9480-3a30316b8e23\") " pod="openstack/swift-proxy-5c7b59dd4f-spxm4" Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.274948 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e743cd5e-e8c6-4fe2-9480-3a30316b8e23-public-tls-certs\") pod \"swift-proxy-5c7b59dd4f-spxm4\" (UID: \"e743cd5e-e8c6-4fe2-9480-3a30316b8e23\") " pod="openstack/swift-proxy-5c7b59dd4f-spxm4" Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.280120 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e743cd5e-e8c6-4fe2-9480-3a30316b8e23-etc-swift\") pod \"swift-proxy-5c7b59dd4f-spxm4\" (UID: \"e743cd5e-e8c6-4fe2-9480-3a30316b8e23\") " pod="openstack/swift-proxy-5c7b59dd4f-spxm4" Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.282800 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4dng\" (UniqueName: \"kubernetes.io/projected/e743cd5e-e8c6-4fe2-9480-3a30316b8e23-kube-api-access-d4dng\") pod \"swift-proxy-5c7b59dd4f-spxm4\" (UID: \"e743cd5e-e8c6-4fe2-9480-3a30316b8e23\") " pod="openstack/swift-proxy-5c7b59dd4f-spxm4" Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.368685 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5c7b59dd4f-spxm4" Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.444936 4787 generic.go:334] "Generic (PLEG): container finished" podID="b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c" containerID="98c33f6c04008cd30bdef34a1de003ea58ef86bd06cb44398a6e45d9e7916bc5" exitCode=0 Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.445011 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c","Type":"ContainerDied","Data":"98c33f6c04008cd30bdef34a1de003ea58ef86bd06cb44398a6e45d9e7916bc5"} Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.456315 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.456564 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="1a020010-9967-4da2-9f0e-6ec5fc461137" containerName="glance-log" containerID="cri-o://ddc800b2f9d11dcc930a8b39989f2bbc9f97a490a9ba0389ce5ceab59fe9641b" gracePeriod=30 Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.456999 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="1a020010-9967-4da2-9f0e-6ec5fc461137" containerName="glance-httpd" containerID="cri-o://c1bdee63e5aab6c9222b16cd9ff592390efcad78109917777cf0033f2a45414c" gracePeriod=30 Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.457438 4787 generic.go:334] "Generic (PLEG): container finished" podID="6808314c-0e06-4fa4-9c3e-269e5952fd3e" containerID="25a024f9e9cad11ef9f8ce3dd6a938d6dba2bec6ef4e26686340150ab47e10f9" exitCode=0 Oct 01 09:53:13 crc kubenswrapper[4787]: I1001 09:53:13.457551 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b69d7b58f-t9lc6" event={"ID":"6808314c-0e06-4fa4-9c3e-269e5952fd3e","Type":"ContainerDied","Data":"25a024f9e9cad11ef9f8ce3dd6a938d6dba2bec6ef4e26686340150ab47e10f9"} Oct 01 09:53:14 crc kubenswrapper[4787]: I1001 09:53:14.510222 4787 generic.go:334] "Generic (PLEG): container finished" podID="b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c" containerID="974cfa14c1d8f423c33e89b90de41d56b73acf24136c57d8454283e7000a8425" exitCode=0 Oct 01 09:53:14 crc kubenswrapper[4787]: I1001 09:53:14.510672 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c","Type":"ContainerDied","Data":"974cfa14c1d8f423c33e89b90de41d56b73acf24136c57d8454283e7000a8425"} Oct 01 09:53:14 crc kubenswrapper[4787]: I1001 09:53:14.515541 4787 generic.go:334] "Generic (PLEG): container finished" podID="1a020010-9967-4da2-9f0e-6ec5fc461137" containerID="ddc800b2f9d11dcc930a8b39989f2bbc9f97a490a9ba0389ce5ceab59fe9641b" exitCode=143 Oct 01 09:53:14 crc kubenswrapper[4787]: I1001 09:53:14.515614 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1a020010-9967-4da2-9f0e-6ec5fc461137","Type":"ContainerDied","Data":"ddc800b2f9d11dcc930a8b39989f2bbc9f97a490a9ba0389ce5ceab59fe9641b"} Oct 01 09:53:14 crc kubenswrapper[4787]: I1001 09:53:14.853740 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-79b69df69b-mfcjt" Oct 01 09:53:14 crc kubenswrapper[4787]: I1001 09:53:14.889544 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-79b69df69b-mfcjt" Oct 01 09:53:16 crc kubenswrapper[4787]: I1001 09:53:16.129821 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:53:16 crc kubenswrapper[4787]: I1001 09:53:16.130464 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="527b1b7d-df6a-4d55-8aba-2f1333cd5534" containerName="glance-log" containerID="cri-o://0fe7ee68c24d07552d2277601f96ad684aedcf7f34ffbb273dc8b5ea7a7d2da8" gracePeriod=30 Oct 01 09:53:16 crc kubenswrapper[4787]: I1001 09:53:16.130610 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="527b1b7d-df6a-4d55-8aba-2f1333cd5534" containerName="glance-httpd" containerID="cri-o://0875d6598ced04ae5adf41e252c6f3988626ac39ce64ccae48a815ad87ee8791" gracePeriod=30 Oct 01 09:53:16 crc kubenswrapper[4787]: I1001 09:53:16.431904 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6b69d7b58f-t9lc6" podUID="6808314c-0e06-4fa4-9c3e-269e5952fd3e" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.154:5353: connect: connection refused" Oct 01 09:53:16 crc kubenswrapper[4787]: I1001 09:53:16.553499 4787 generic.go:334] "Generic (PLEG): container finished" podID="527b1b7d-df6a-4d55-8aba-2f1333cd5534" containerID="0fe7ee68c24d07552d2277601f96ad684aedcf7f34ffbb273dc8b5ea7a7d2da8" exitCode=143 Oct 01 09:53:16 crc kubenswrapper[4787]: I1001 09:53:16.553563 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"527b1b7d-df6a-4d55-8aba-2f1333cd5534","Type":"ContainerDied","Data":"0fe7ee68c24d07552d2277601f96ad684aedcf7f34ffbb273dc8b5ea7a7d2da8"} Oct 01 09:53:17 crc kubenswrapper[4787]: I1001 09:53:17.291756 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-887c78f44-f74pm" Oct 01 09:53:17 crc kubenswrapper[4787]: I1001 09:53:17.433384 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-887c78f44-f74pm" Oct 01 09:53:17 crc kubenswrapper[4787]: I1001 09:53:17.521144 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-nhmnw"] Oct 01 09:53:17 crc kubenswrapper[4787]: I1001 09:53:17.529244 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-79b69df69b-mfcjt"] Oct 01 09:53:17 crc kubenswrapper[4787]: I1001 09:53:17.529408 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-nhmnw" Oct 01 09:53:17 crc kubenswrapper[4787]: I1001 09:53:17.529544 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-79b69df69b-mfcjt" podUID="0244e33a-702c-412d-8074-a2c910f375d0" containerName="barbican-api-log" containerID="cri-o://f6e497aeeaf590b497d72aedd95d80d7aecc82ff93802c4e050cb6ed19e5e13c" gracePeriod=30 Oct 01 09:53:17 crc kubenswrapper[4787]: I1001 09:53:17.529689 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-79b69df69b-mfcjt" podUID="0244e33a-702c-412d-8074-a2c910f375d0" containerName="barbican-api" containerID="cri-o://9809e8583584726a8893b9ddf2134ade7d5b049f4935bc46a74fc053a250125d" gracePeriod=30 Oct 01 09:53:17 crc kubenswrapper[4787]: I1001 09:53:17.551517 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-nhmnw"] Oct 01 09:53:17 crc kubenswrapper[4787]: I1001 09:53:17.590918 4787 generic.go:334] "Generic (PLEG): container finished" podID="1a020010-9967-4da2-9f0e-6ec5fc461137" containerID="c1bdee63e5aab6c9222b16cd9ff592390efcad78109917777cf0033f2a45414c" exitCode=0 Oct 01 09:53:17 crc kubenswrapper[4787]: I1001 09:53:17.591035 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1a020010-9967-4da2-9f0e-6ec5fc461137","Type":"ContainerDied","Data":"c1bdee63e5aab6c9222b16cd9ff592390efcad78109917777cf0033f2a45414c"} Oct 01 09:53:17 crc kubenswrapper[4787]: I1001 09:53:17.606621 4787 generic.go:334] "Generic (PLEG): container finished" podID="ca8b9373-e5e5-4dc1-8995-44ae47abbb13" containerID="636e18451477b045320e69118c705fe6303e8717966fab13f8b4ae36f44898f5" exitCode=0 Oct 01 09:53:17 crc kubenswrapper[4787]: I1001 09:53:17.607715 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-8mrcw" event={"ID":"ca8b9373-e5e5-4dc1-8995-44ae47abbb13","Type":"ContainerDied","Data":"636e18451477b045320e69118c705fe6303e8717966fab13f8b4ae36f44898f5"} Oct 01 09:53:17 crc kubenswrapper[4787]: I1001 09:53:17.654955 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vxnh\" (UniqueName: \"kubernetes.io/projected/931c8c97-57b0-4b0c-a97d-498eaad48543-kube-api-access-7vxnh\") pod \"nova-api-db-create-nhmnw\" (UID: \"931c8c97-57b0-4b0c-a97d-498eaad48543\") " pod="openstack/nova-api-db-create-nhmnw" Oct 01 09:53:17 crc kubenswrapper[4787]: I1001 09:53:17.756576 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vxnh\" (UniqueName: \"kubernetes.io/projected/931c8c97-57b0-4b0c-a97d-498eaad48543-kube-api-access-7vxnh\") pod \"nova-api-db-create-nhmnw\" (UID: \"931c8c97-57b0-4b0c-a97d-498eaad48543\") " pod="openstack/nova-api-db-create-nhmnw" Oct 01 09:53:17 crc kubenswrapper[4787]: I1001 09:53:17.818999 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-69qw2"] Oct 01 09:53:17 crc kubenswrapper[4787]: I1001 09:53:17.821286 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-69qw2" Oct 01 09:53:17 crc kubenswrapper[4787]: I1001 09:53:17.825144 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vxnh\" (UniqueName: \"kubernetes.io/projected/931c8c97-57b0-4b0c-a97d-498eaad48543-kube-api-access-7vxnh\") pod \"nova-api-db-create-nhmnw\" (UID: \"931c8c97-57b0-4b0c-a97d-498eaad48543\") " pod="openstack/nova-api-db-create-nhmnw" Oct 01 09:53:17 crc kubenswrapper[4787]: I1001 09:53:17.854902 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-69qw2"] Oct 01 09:53:17 crc kubenswrapper[4787]: I1001 09:53:17.867932 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-nhmnw" Oct 01 09:53:17 crc kubenswrapper[4787]: I1001 09:53:17.911607 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-9kp59"] Oct 01 09:53:17 crc kubenswrapper[4787]: I1001 09:53:17.913165 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-9kp59" Oct 01 09:53:17 crc kubenswrapper[4787]: I1001 09:53:17.926981 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-9kp59"] Oct 01 09:53:17 crc kubenswrapper[4787]: I1001 09:53:17.964693 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sd2w\" (UniqueName: \"kubernetes.io/projected/66351c20-c4db-489c-857f-c82107d2be61-kube-api-access-4sd2w\") pod \"nova-cell0-db-create-69qw2\" (UID: \"66351c20-c4db-489c-857f-c82107d2be61\") " pod="openstack/nova-cell0-db-create-69qw2" Oct 01 09:53:17 crc kubenswrapper[4787]: I1001 09:53:17.964838 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kpmn\" (UniqueName: \"kubernetes.io/projected/a7ba3660-dca6-4843-b560-efd0ea158258-kube-api-access-6kpmn\") pod \"nova-cell1-db-create-9kp59\" (UID: \"a7ba3660-dca6-4843-b560-efd0ea158258\") " pod="openstack/nova-cell1-db-create-9kp59" Oct 01 09:53:18 crc kubenswrapper[4787]: I1001 09:53:18.066791 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sd2w\" (UniqueName: \"kubernetes.io/projected/66351c20-c4db-489c-857f-c82107d2be61-kube-api-access-4sd2w\") pod \"nova-cell0-db-create-69qw2\" (UID: \"66351c20-c4db-489c-857f-c82107d2be61\") " pod="openstack/nova-cell0-db-create-69qw2" Oct 01 09:53:18 crc kubenswrapper[4787]: I1001 09:53:18.066867 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kpmn\" (UniqueName: \"kubernetes.io/projected/a7ba3660-dca6-4843-b560-efd0ea158258-kube-api-access-6kpmn\") pod \"nova-cell1-db-create-9kp59\" (UID: \"a7ba3660-dca6-4843-b560-efd0ea158258\") " pod="openstack/nova-cell1-db-create-9kp59" Oct 01 09:53:18 crc kubenswrapper[4787]: I1001 09:53:18.091264 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sd2w\" (UniqueName: \"kubernetes.io/projected/66351c20-c4db-489c-857f-c82107d2be61-kube-api-access-4sd2w\") pod \"nova-cell0-db-create-69qw2\" (UID: \"66351c20-c4db-489c-857f-c82107d2be61\") " pod="openstack/nova-cell0-db-create-69qw2" Oct 01 09:53:18 crc kubenswrapper[4787]: I1001 09:53:18.093413 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kpmn\" (UniqueName: \"kubernetes.io/projected/a7ba3660-dca6-4843-b560-efd0ea158258-kube-api-access-6kpmn\") pod \"nova-cell1-db-create-9kp59\" (UID: \"a7ba3660-dca6-4843-b560-efd0ea158258\") " pod="openstack/nova-cell1-db-create-9kp59" Oct 01 09:53:18 crc kubenswrapper[4787]: I1001 09:53:18.169421 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-69qw2" Oct 01 09:53:18 crc kubenswrapper[4787]: I1001 09:53:18.242575 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-9kp59" Oct 01 09:53:18 crc kubenswrapper[4787]: I1001 09:53:18.623924 4787 generic.go:334] "Generic (PLEG): container finished" podID="0244e33a-702c-412d-8074-a2c910f375d0" containerID="f6e497aeeaf590b497d72aedd95d80d7aecc82ff93802c4e050cb6ed19e5e13c" exitCode=143 Oct 01 09:53:18 crc kubenswrapper[4787]: I1001 09:53:18.624295 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-79b69df69b-mfcjt" event={"ID":"0244e33a-702c-412d-8074-a2c910f375d0","Type":"ContainerDied","Data":"f6e497aeeaf590b497d72aedd95d80d7aecc82ff93802c4e050cb6ed19e5e13c"} Oct 01 09:53:19 crc kubenswrapper[4787]: I1001 09:53:19.641406 4787 generic.go:334] "Generic (PLEG): container finished" podID="527b1b7d-df6a-4d55-8aba-2f1333cd5534" containerID="0875d6598ced04ae5adf41e252c6f3988626ac39ce64ccae48a815ad87ee8791" exitCode=0 Oct 01 09:53:19 crc kubenswrapper[4787]: I1001 09:53:19.641456 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"527b1b7d-df6a-4d55-8aba-2f1333cd5534","Type":"ContainerDied","Data":"0875d6598ced04ae5adf41e252c6f3988626ac39ce64ccae48a815ad87ee8791"} Oct 01 09:53:20 crc kubenswrapper[4787]: I1001 09:53:20.737333 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-79b69df69b-mfcjt" podUID="0244e33a-702c-412d-8074-a2c910f375d0" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.163:9311/healthcheck\": read tcp 10.217.0.2:36664->10.217.0.163:9311: read: connection reset by peer" Oct 01 09:53:20 crc kubenswrapper[4787]: I1001 09:53:20.737926 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-79b69df69b-mfcjt" podUID="0244e33a-702c-412d-8074-a2c910f375d0" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.163:9311/healthcheck\": read tcp 10.217.0.2:36656->10.217.0.163:9311: read: connection reset by peer" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.356363 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-8mrcw" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.439469 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca8b9373-e5e5-4dc1-8995-44ae47abbb13-scripts\") pod \"ca8b9373-e5e5-4dc1-8995-44ae47abbb13\" (UID: \"ca8b9373-e5e5-4dc1-8995-44ae47abbb13\") " Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.440142 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ca8b9373-e5e5-4dc1-8995-44ae47abbb13-etc-machine-id\") pod \"ca8b9373-e5e5-4dc1-8995-44ae47abbb13\" (UID: \"ca8b9373-e5e5-4dc1-8995-44ae47abbb13\") " Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.440222 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca8b9373-e5e5-4dc1-8995-44ae47abbb13-config-data\") pod \"ca8b9373-e5e5-4dc1-8995-44ae47abbb13\" (UID: \"ca8b9373-e5e5-4dc1-8995-44ae47abbb13\") " Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.440277 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6kfv2\" (UniqueName: \"kubernetes.io/projected/ca8b9373-e5e5-4dc1-8995-44ae47abbb13-kube-api-access-6kfv2\") pod \"ca8b9373-e5e5-4dc1-8995-44ae47abbb13\" (UID: \"ca8b9373-e5e5-4dc1-8995-44ae47abbb13\") " Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.440312 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ca8b9373-e5e5-4dc1-8995-44ae47abbb13-db-sync-config-data\") pod \"ca8b9373-e5e5-4dc1-8995-44ae47abbb13\" (UID: \"ca8b9373-e5e5-4dc1-8995-44ae47abbb13\") " Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.440456 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca8b9373-e5e5-4dc1-8995-44ae47abbb13-combined-ca-bundle\") pod \"ca8b9373-e5e5-4dc1-8995-44ae47abbb13\" (UID: \"ca8b9373-e5e5-4dc1-8995-44ae47abbb13\") " Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.442787 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ca8b9373-e5e5-4dc1-8995-44ae47abbb13-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ca8b9373-e5e5-4dc1-8995-44ae47abbb13" (UID: "ca8b9373-e5e5-4dc1-8995-44ae47abbb13"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.450839 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca8b9373-e5e5-4dc1-8995-44ae47abbb13-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "ca8b9373-e5e5-4dc1-8995-44ae47abbb13" (UID: "ca8b9373-e5e5-4dc1-8995-44ae47abbb13"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.450939 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca8b9373-e5e5-4dc1-8995-44ae47abbb13-kube-api-access-6kfv2" (OuterVolumeSpecName: "kube-api-access-6kfv2") pod "ca8b9373-e5e5-4dc1-8995-44ae47abbb13" (UID: "ca8b9373-e5e5-4dc1-8995-44ae47abbb13"). InnerVolumeSpecName "kube-api-access-6kfv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.451169 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca8b9373-e5e5-4dc1-8995-44ae47abbb13-scripts" (OuterVolumeSpecName: "scripts") pod "ca8b9373-e5e5-4dc1-8995-44ae47abbb13" (UID: "ca8b9373-e5e5-4dc1-8995-44ae47abbb13"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.509392 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca8b9373-e5e5-4dc1-8995-44ae47abbb13-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ca8b9373-e5e5-4dc1-8995-44ae47abbb13" (UID: "ca8b9373-e5e5-4dc1-8995-44ae47abbb13"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.542648 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca8b9373-e5e5-4dc1-8995-44ae47abbb13-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.543508 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca8b9373-e5e5-4dc1-8995-44ae47abbb13-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.543524 4787 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ca8b9373-e5e5-4dc1-8995-44ae47abbb13-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.543534 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6kfv2\" (UniqueName: \"kubernetes.io/projected/ca8b9373-e5e5-4dc1-8995-44ae47abbb13-kube-api-access-6kfv2\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.543545 4787 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ca8b9373-e5e5-4dc1-8995-44ae47abbb13-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.580055 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca8b9373-e5e5-4dc1-8995-44ae47abbb13-config-data" (OuterVolumeSpecName: "config-data") pod "ca8b9373-e5e5-4dc1-8995-44ae47abbb13" (UID: "ca8b9373-e5e5-4dc1-8995-44ae47abbb13"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.628227 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b69d7b58f-t9lc6" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.645725 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca8b9373-e5e5-4dc1-8995-44ae47abbb13-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.681346 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-8mrcw" event={"ID":"ca8b9373-e5e5-4dc1-8995-44ae47abbb13","Type":"ContainerDied","Data":"7dd24c6a97a49814118f912a32a5c7e3c9a96581d698f74375becf0222caa493"} Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.681400 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7dd24c6a97a49814118f912a32a5c7e3c9a96581d698f74375becf0222caa493" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.681490 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-8mrcw" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.693426 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b69d7b58f-t9lc6" event={"ID":"6808314c-0e06-4fa4-9c3e-269e5952fd3e","Type":"ContainerDied","Data":"199c7ef098dbbfd089ac3178bd69c016445a7ec524defee9b244ef53ba6497b7"} Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.693509 4787 scope.go:117] "RemoveContainer" containerID="25a024f9e9cad11ef9f8ce3dd6a938d6dba2bec6ef4e26686340150ab47e10f9" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.693451 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b69d7b58f-t9lc6" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.711177 4787 generic.go:334] "Generic (PLEG): container finished" podID="0244e33a-702c-412d-8074-a2c910f375d0" containerID="9809e8583584726a8893b9ddf2134ade7d5b049f4935bc46a74fc053a250125d" exitCode=0 Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.711271 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-79b69df69b-mfcjt" event={"ID":"0244e33a-702c-412d-8074-a2c910f375d0","Type":"ContainerDied","Data":"9809e8583584726a8893b9ddf2134ade7d5b049f4935bc46a74fc053a250125d"} Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.743825 4787 scope.go:117] "RemoveContainer" containerID="7871e8dfa99a35696ecc214ea9f6b3f9e404c4b4ca98c24c34741b14a1fb92a1" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.747920 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rcfg4\" (UniqueName: \"kubernetes.io/projected/6808314c-0e06-4fa4-9c3e-269e5952fd3e-kube-api-access-rcfg4\") pod \"6808314c-0e06-4fa4-9c3e-269e5952fd3e\" (UID: \"6808314c-0e06-4fa4-9c3e-269e5952fd3e\") " Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.748156 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6808314c-0e06-4fa4-9c3e-269e5952fd3e-ovsdbserver-sb\") pod \"6808314c-0e06-4fa4-9c3e-269e5952fd3e\" (UID: \"6808314c-0e06-4fa4-9c3e-269e5952fd3e\") " Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.748264 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6808314c-0e06-4fa4-9c3e-269e5952fd3e-dns-svc\") pod \"6808314c-0e06-4fa4-9c3e-269e5952fd3e\" (UID: \"6808314c-0e06-4fa4-9c3e-269e5952fd3e\") " Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.748741 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6808314c-0e06-4fa4-9c3e-269e5952fd3e-ovsdbserver-nb\") pod \"6808314c-0e06-4fa4-9c3e-269e5952fd3e\" (UID: \"6808314c-0e06-4fa4-9c3e-269e5952fd3e\") " Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.748852 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6808314c-0e06-4fa4-9c3e-269e5952fd3e-config\") pod \"6808314c-0e06-4fa4-9c3e-269e5952fd3e\" (UID: \"6808314c-0e06-4fa4-9c3e-269e5952fd3e\") " Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.748895 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6808314c-0e06-4fa4-9c3e-269e5952fd3e-dns-swift-storage-0\") pod \"6808314c-0e06-4fa4-9c3e-269e5952fd3e\" (UID: \"6808314c-0e06-4fa4-9c3e-269e5952fd3e\") " Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.772536 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6808314c-0e06-4fa4-9c3e-269e5952fd3e-kube-api-access-rcfg4" (OuterVolumeSpecName: "kube-api-access-rcfg4") pod "6808314c-0e06-4fa4-9c3e-269e5952fd3e" (UID: "6808314c-0e06-4fa4-9c3e-269e5952fd3e"). InnerVolumeSpecName "kube-api-access-rcfg4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.822281 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-79b69df69b-mfcjt" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.859263 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rcfg4\" (UniqueName: \"kubernetes.io/projected/6808314c-0e06-4fa4-9c3e-269e5952fd3e-kube-api-access-rcfg4\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.908967 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.910386 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6808314c-0e06-4fa4-9c3e-269e5952fd3e-config" (OuterVolumeSpecName: "config") pod "6808314c-0e06-4fa4-9c3e-269e5952fd3e" (UID: "6808314c-0e06-4fa4-9c3e-269e5952fd3e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.910757 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6808314c-0e06-4fa4-9c3e-269e5952fd3e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6808314c-0e06-4fa4-9c3e-269e5952fd3e" (UID: "6808314c-0e06-4fa4-9c3e-269e5952fd3e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.911513 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.922920 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6808314c-0e06-4fa4-9c3e-269e5952fd3e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6808314c-0e06-4fa4-9c3e-269e5952fd3e" (UID: "6808314c-0e06-4fa4-9c3e-269e5952fd3e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.923178 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6808314c-0e06-4fa4-9c3e-269e5952fd3e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6808314c-0e06-4fa4-9c3e-269e5952fd3e" (UID: "6808314c-0e06-4fa4-9c3e-269e5952fd3e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.941544 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6808314c-0e06-4fa4-9c3e-269e5952fd3e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6808314c-0e06-4fa4-9c3e-269e5952fd3e" (UID: "6808314c-0e06-4fa4-9c3e-269e5952fd3e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.960598 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-config-data\") pod \"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c\" (UID: \"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c\") " Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.960686 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdw8q\" (UniqueName: \"kubernetes.io/projected/0244e33a-702c-412d-8074-a2c910f375d0-kube-api-access-kdw8q\") pod \"0244e33a-702c-412d-8074-a2c910f375d0\" (UID: \"0244e33a-702c-412d-8074-a2c910f375d0\") " Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.960747 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7zvd\" (UniqueName: \"kubernetes.io/projected/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-kube-api-access-j7zvd\") pod \"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c\" (UID: \"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c\") " Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.960786 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0244e33a-702c-412d-8074-a2c910f375d0-combined-ca-bundle\") pod \"0244e33a-702c-412d-8074-a2c910f375d0\" (UID: \"0244e33a-702c-412d-8074-a2c910f375d0\") " Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.960846 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0244e33a-702c-412d-8074-a2c910f375d0-config-data\") pod \"0244e33a-702c-412d-8074-a2c910f375d0\" (UID: \"0244e33a-702c-412d-8074-a2c910f375d0\") " Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.960892 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0244e33a-702c-412d-8074-a2c910f375d0-config-data-custom\") pod \"0244e33a-702c-412d-8074-a2c910f375d0\" (UID: \"0244e33a-702c-412d-8074-a2c910f375d0\") " Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.960914 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0244e33a-702c-412d-8074-a2c910f375d0-logs\") pod \"0244e33a-702c-412d-8074-a2c910f375d0\" (UID: \"0244e33a-702c-412d-8074-a2c910f375d0\") " Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.960946 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-log-httpd\") pod \"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c\" (UID: \"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c\") " Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.960963 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-scripts\") pod \"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c\" (UID: \"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c\") " Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.961013 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-run-httpd\") pod \"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c\" (UID: \"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c\") " Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.961106 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-combined-ca-bundle\") pod \"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c\" (UID: \"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c\") " Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.961129 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-sg-core-conf-yaml\") pod \"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c\" (UID: \"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c\") " Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.961536 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6808314c-0e06-4fa4-9c3e-269e5952fd3e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.961548 4787 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6808314c-0e06-4fa4-9c3e-269e5952fd3e-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.961558 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6808314c-0e06-4fa4-9c3e-269e5952fd3e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.961566 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6808314c-0e06-4fa4-9c3e-269e5952fd3e-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.961575 4787 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6808314c-0e06-4fa4-9c3e-269e5952fd3e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.966858 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0244e33a-702c-412d-8074-a2c910f375d0-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "0244e33a-702c-412d-8074-a2c910f375d0" (UID: "0244e33a-702c-412d-8074-a2c910f375d0"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.973880 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c" (UID: "b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.974759 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0244e33a-702c-412d-8074-a2c910f375d0-logs" (OuterVolumeSpecName: "logs") pod "0244e33a-702c-412d-8074-a2c910f375d0" (UID: "0244e33a-702c-412d-8074-a2c910f375d0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.975152 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-kube-api-access-j7zvd" (OuterVolumeSpecName: "kube-api-access-j7zvd") pod "b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c" (UID: "b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c"). InnerVolumeSpecName "kube-api-access-j7zvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.978508 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0244e33a-702c-412d-8074-a2c910f375d0-kube-api-access-kdw8q" (OuterVolumeSpecName: "kube-api-access-kdw8q") pod "0244e33a-702c-412d-8074-a2c910f375d0" (UID: "0244e33a-702c-412d-8074-a2c910f375d0"). InnerVolumeSpecName "kube-api-access-kdw8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.978684 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-scripts" (OuterVolumeSpecName: "scripts") pod "b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c" (UID: "b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:21 crc kubenswrapper[4787]: I1001 09:53:21.978782 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c" (UID: "b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.004830 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c" (UID: "b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.027411 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0244e33a-702c-412d-8074-a2c910f375d0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0244e33a-702c-412d-8074-a2c910f375d0" (UID: "0244e33a-702c-412d-8074-a2c910f375d0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.032232 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0244e33a-702c-412d-8074-a2c910f375d0-config-data" (OuterVolumeSpecName: "config-data") pod "0244e33a-702c-412d-8074-a2c910f375d0" (UID: "0244e33a-702c-412d-8074-a2c910f375d0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.033196 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b69d7b58f-t9lc6"] Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.038204 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.041253 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b69d7b58f-t9lc6"] Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.064821 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"1a020010-9967-4da2-9f0e-6ec5fc461137\" (UID: \"1a020010-9967-4da2-9f0e-6ec5fc461137\") " Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.064965 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1a020010-9967-4da2-9f0e-6ec5fc461137-httpd-run\") pod \"1a020010-9967-4da2-9f0e-6ec5fc461137\" (UID: \"1a020010-9967-4da2-9f0e-6ec5fc461137\") " Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.064987 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a020010-9967-4da2-9f0e-6ec5fc461137-combined-ca-bundle\") pod \"1a020010-9967-4da2-9f0e-6ec5fc461137\" (UID: \"1a020010-9967-4da2-9f0e-6ec5fc461137\") " Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.065069 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvtc\" (UniqueName: \"kubernetes.io/projected/1a020010-9967-4da2-9f0e-6ec5fc461137-kube-api-access-zkvtc\") pod \"1a020010-9967-4da2-9f0e-6ec5fc461137\" (UID: \"1a020010-9967-4da2-9f0e-6ec5fc461137\") " Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.065125 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a020010-9967-4da2-9f0e-6ec5fc461137-public-tls-certs\") pod \"1a020010-9967-4da2-9f0e-6ec5fc461137\" (UID: \"1a020010-9967-4da2-9f0e-6ec5fc461137\") " Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.065168 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a020010-9967-4da2-9f0e-6ec5fc461137-scripts\") pod \"1a020010-9967-4da2-9f0e-6ec5fc461137\" (UID: \"1a020010-9967-4da2-9f0e-6ec5fc461137\") " Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.065200 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a020010-9967-4da2-9f0e-6ec5fc461137-logs\") pod \"1a020010-9967-4da2-9f0e-6ec5fc461137\" (UID: \"1a020010-9967-4da2-9f0e-6ec5fc461137\") " Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.065223 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a020010-9967-4da2-9f0e-6ec5fc461137-config-data\") pod \"1a020010-9967-4da2-9f0e-6ec5fc461137\" (UID: \"1a020010-9967-4da2-9f0e-6ec5fc461137\") " Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.065564 4787 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.065576 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdw8q\" (UniqueName: \"kubernetes.io/projected/0244e33a-702c-412d-8074-a2c910f375d0-kube-api-access-kdw8q\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.065587 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7zvd\" (UniqueName: \"kubernetes.io/projected/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-kube-api-access-j7zvd\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.065597 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0244e33a-702c-412d-8074-a2c910f375d0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.065606 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0244e33a-702c-412d-8074-a2c910f375d0-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.065614 4787 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0244e33a-702c-412d-8074-a2c910f375d0-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.065621 4787 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0244e33a-702c-412d-8074-a2c910f375d0-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.065629 4787 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.065637 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.065645 4787 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.066237 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a020010-9967-4da2-9f0e-6ec5fc461137-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "1a020010-9967-4da2-9f0e-6ec5fc461137" (UID: "1a020010-9967-4da2-9f0e-6ec5fc461137"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.067022 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a020010-9967-4da2-9f0e-6ec5fc461137-logs" (OuterVolumeSpecName: "logs") pod "1a020010-9967-4da2-9f0e-6ec5fc461137" (UID: "1a020010-9967-4da2-9f0e-6ec5fc461137"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.071590 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a020010-9967-4da2-9f0e-6ec5fc461137-kube-api-access-zkvtc" (OuterVolumeSpecName: "kube-api-access-zkvtc") pod "1a020010-9967-4da2-9f0e-6ec5fc461137" (UID: "1a020010-9967-4da2-9f0e-6ec5fc461137"). InnerVolumeSpecName "kube-api-access-zkvtc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.071814 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "1a020010-9967-4da2-9f0e-6ec5fc461137" (UID: "1a020010-9967-4da2-9f0e-6ec5fc461137"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.074675 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a020010-9967-4da2-9f0e-6ec5fc461137-scripts" (OuterVolumeSpecName: "scripts") pod "1a020010-9967-4da2-9f0e-6ec5fc461137" (UID: "1a020010-9967-4da2-9f0e-6ec5fc461137"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.091463 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c" (UID: "b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.108504 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-config-data" (OuterVolumeSpecName: "config-data") pod "b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c" (UID: "b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.122777 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a020010-9967-4da2-9f0e-6ec5fc461137-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1a020010-9967-4da2-9f0e-6ec5fc461137" (UID: "1a020010-9967-4da2-9f0e-6ec5fc461137"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.126001 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a020010-9967-4da2-9f0e-6ec5fc461137-config-data" (OuterVolumeSpecName: "config-data") pod "1a020010-9967-4da2-9f0e-6ec5fc461137" (UID: "1a020010-9967-4da2-9f0e-6ec5fc461137"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.140023 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a020010-9967-4da2-9f0e-6ec5fc461137-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "1a020010-9967-4da2-9f0e-6ec5fc461137" (UID: "1a020010-9967-4da2-9f0e-6ec5fc461137"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.167805 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/527b1b7d-df6a-4d55-8aba-2f1333cd5534-config-data\") pod \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\" (UID: \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\") " Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.168017 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/527b1b7d-df6a-4d55-8aba-2f1333cd5534-internal-tls-certs\") pod \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\" (UID: \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\") " Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.168144 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/527b1b7d-df6a-4d55-8aba-2f1333cd5534-logs\") pod \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\" (UID: \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\") " Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.168171 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\" (UID: \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\") " Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.168218 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmhwl\" (UniqueName: \"kubernetes.io/projected/527b1b7d-df6a-4d55-8aba-2f1333cd5534-kube-api-access-cmhwl\") pod \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\" (UID: \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\") " Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.168250 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/527b1b7d-df6a-4d55-8aba-2f1333cd5534-combined-ca-bundle\") pod \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\" (UID: \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\") " Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.169226 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/527b1b7d-df6a-4d55-8aba-2f1333cd5534-scripts\") pod \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\" (UID: \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\") " Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.169259 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/527b1b7d-df6a-4d55-8aba-2f1333cd5534-httpd-run\") pod \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\" (UID: \"527b1b7d-df6a-4d55-8aba-2f1333cd5534\") " Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.170782 4787 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1a020010-9967-4da2-9f0e-6ec5fc461137-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.170811 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a020010-9967-4da2-9f0e-6ec5fc461137-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.170830 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvtc\" (UniqueName: \"kubernetes.io/projected/1a020010-9967-4da2-9f0e-6ec5fc461137-kube-api-access-zkvtc\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.170841 4787 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a020010-9967-4da2-9f0e-6ec5fc461137-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.170852 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a020010-9967-4da2-9f0e-6ec5fc461137-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.170863 4787 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a020010-9967-4da2-9f0e-6ec5fc461137-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.170873 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a020010-9967-4da2-9f0e-6ec5fc461137-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.170902 4787 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.170916 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.170928 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.177291 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/527b1b7d-df6a-4d55-8aba-2f1333cd5534-logs" (OuterVolumeSpecName: "logs") pod "527b1b7d-df6a-4d55-8aba-2f1333cd5534" (UID: "527b1b7d-df6a-4d55-8aba-2f1333cd5534"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.177833 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/527b1b7d-df6a-4d55-8aba-2f1333cd5534-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "527b1b7d-df6a-4d55-8aba-2f1333cd5534" (UID: "527b1b7d-df6a-4d55-8aba-2f1333cd5534"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.178921 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/527b1b7d-df6a-4d55-8aba-2f1333cd5534-kube-api-access-cmhwl" (OuterVolumeSpecName: "kube-api-access-cmhwl") pod "527b1b7d-df6a-4d55-8aba-2f1333cd5534" (UID: "527b1b7d-df6a-4d55-8aba-2f1333cd5534"). InnerVolumeSpecName "kube-api-access-cmhwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.180841 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "527b1b7d-df6a-4d55-8aba-2f1333cd5534" (UID: "527b1b7d-df6a-4d55-8aba-2f1333cd5534"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.186050 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/527b1b7d-df6a-4d55-8aba-2f1333cd5534-scripts" (OuterVolumeSpecName: "scripts") pod "527b1b7d-df6a-4d55-8aba-2f1333cd5534" (UID: "527b1b7d-df6a-4d55-8aba-2f1333cd5534"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.216317 4787 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.225804 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/527b1b7d-df6a-4d55-8aba-2f1333cd5534-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "527b1b7d-df6a-4d55-8aba-2f1333cd5534" (UID: "527b1b7d-df6a-4d55-8aba-2f1333cd5534"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.260247 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/527b1b7d-df6a-4d55-8aba-2f1333cd5534-config-data" (OuterVolumeSpecName: "config-data") pod "527b1b7d-df6a-4d55-8aba-2f1333cd5534" (UID: "527b1b7d-df6a-4d55-8aba-2f1333cd5534"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.265463 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/527b1b7d-df6a-4d55-8aba-2f1333cd5534-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "527b1b7d-df6a-4d55-8aba-2f1333cd5534" (UID: "527b1b7d-df6a-4d55-8aba-2f1333cd5534"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.272626 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/527b1b7d-df6a-4d55-8aba-2f1333cd5534-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.272660 4787 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/527b1b7d-df6a-4d55-8aba-2f1333cd5534-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.272673 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/527b1b7d-df6a-4d55-8aba-2f1333cd5534-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.272682 4787 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/527b1b7d-df6a-4d55-8aba-2f1333cd5534-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.272693 4787 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/527b1b7d-df6a-4d55-8aba-2f1333cd5534-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.272727 4787 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.272737 4787 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.272748 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmhwl\" (UniqueName: \"kubernetes.io/projected/527b1b7d-df6a-4d55-8aba-2f1333cd5534-kube-api-access-cmhwl\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.272757 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/527b1b7d-df6a-4d55-8aba-2f1333cd5534-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.297414 4787 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.374049 4787 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.390147 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-9kp59"] Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.422184 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-69qw2"] Oct 01 09:53:22 crc kubenswrapper[4787]: W1001 09:53:22.430756 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod931c8c97_57b0_4b0c_a97d_498eaad48543.slice/crio-b392312040e6c776e604a490758bc2c11eb31fe0e20e065a9609db1d7ed3dd1f WatchSource:0}: Error finding container b392312040e6c776e604a490758bc2c11eb31fe0e20e065a9609db1d7ed3dd1f: Status 404 returned error can't find the container with id b392312040e6c776e604a490758bc2c11eb31fe0e20e065a9609db1d7ed3dd1f Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.456543 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-nhmnw"] Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.493780 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5c7b59dd4f-spxm4"] Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.536356 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6808314c-0e06-4fa4-9c3e-269e5952fd3e" path="/var/lib/kubelet/pods/6808314c-0e06-4fa4-9c3e-269e5952fd3e/volumes" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.657990 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 09:53:22 crc kubenswrapper[4787]: E1001 09:53:22.658689 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c" containerName="ceilometer-notification-agent" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.658707 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c" containerName="ceilometer-notification-agent" Oct 01 09:53:22 crc kubenswrapper[4787]: E1001 09:53:22.658718 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="527b1b7d-df6a-4d55-8aba-2f1333cd5534" containerName="glance-log" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.658724 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="527b1b7d-df6a-4d55-8aba-2f1333cd5534" containerName="glance-log" Oct 01 09:53:22 crc kubenswrapper[4787]: E1001 09:53:22.658735 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6808314c-0e06-4fa4-9c3e-269e5952fd3e" containerName="dnsmasq-dns" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.658741 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="6808314c-0e06-4fa4-9c3e-269e5952fd3e" containerName="dnsmasq-dns" Oct 01 09:53:22 crc kubenswrapper[4787]: E1001 09:53:22.658757 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c" containerName="proxy-httpd" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.658763 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c" containerName="proxy-httpd" Oct 01 09:53:22 crc kubenswrapper[4787]: E1001 09:53:22.658775 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c" containerName="sg-core" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.658781 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c" containerName="sg-core" Oct 01 09:53:22 crc kubenswrapper[4787]: E1001 09:53:22.658794 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a020010-9967-4da2-9f0e-6ec5fc461137" containerName="glance-log" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.658801 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a020010-9967-4da2-9f0e-6ec5fc461137" containerName="glance-log" Oct 01 09:53:22 crc kubenswrapper[4787]: E1001 09:53:22.658813 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="527b1b7d-df6a-4d55-8aba-2f1333cd5534" containerName="glance-httpd" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.658819 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="527b1b7d-df6a-4d55-8aba-2f1333cd5534" containerName="glance-httpd" Oct 01 09:53:22 crc kubenswrapper[4787]: E1001 09:53:22.658827 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0244e33a-702c-412d-8074-a2c910f375d0" containerName="barbican-api" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.658832 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="0244e33a-702c-412d-8074-a2c910f375d0" containerName="barbican-api" Oct 01 09:53:22 crc kubenswrapper[4787]: E1001 09:53:22.658840 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c" containerName="ceilometer-central-agent" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.658847 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c" containerName="ceilometer-central-agent" Oct 01 09:53:22 crc kubenswrapper[4787]: E1001 09:53:22.658859 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0244e33a-702c-412d-8074-a2c910f375d0" containerName="barbican-api-log" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.658865 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="0244e33a-702c-412d-8074-a2c910f375d0" containerName="barbican-api-log" Oct 01 09:53:22 crc kubenswrapper[4787]: E1001 09:53:22.658876 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6808314c-0e06-4fa4-9c3e-269e5952fd3e" containerName="init" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.658883 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="6808314c-0e06-4fa4-9c3e-269e5952fd3e" containerName="init" Oct 01 09:53:22 crc kubenswrapper[4787]: E1001 09:53:22.658897 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a020010-9967-4da2-9f0e-6ec5fc461137" containerName="glance-httpd" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.658903 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a020010-9967-4da2-9f0e-6ec5fc461137" containerName="glance-httpd" Oct 01 09:53:22 crc kubenswrapper[4787]: E1001 09:53:22.658913 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca8b9373-e5e5-4dc1-8995-44ae47abbb13" containerName="cinder-db-sync" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.658920 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca8b9373-e5e5-4dc1-8995-44ae47abbb13" containerName="cinder-db-sync" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.662436 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="0244e33a-702c-412d-8074-a2c910f375d0" containerName="barbican-api-log" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.662473 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c" containerName="ceilometer-central-agent" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.662485 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="527b1b7d-df6a-4d55-8aba-2f1333cd5534" containerName="glance-log" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.662500 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c" containerName="ceilometer-notification-agent" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.662509 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a020010-9967-4da2-9f0e-6ec5fc461137" containerName="glance-httpd" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.662519 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="6808314c-0e06-4fa4-9c3e-269e5952fd3e" containerName="dnsmasq-dns" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.662529 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="527b1b7d-df6a-4d55-8aba-2f1333cd5534" containerName="glance-httpd" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.662542 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="0244e33a-702c-412d-8074-a2c910f375d0" containerName="barbican-api" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.662551 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c" containerName="proxy-httpd" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.662559 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c" containerName="sg-core" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.662567 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca8b9373-e5e5-4dc1-8995-44ae47abbb13" containerName="cinder-db-sync" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.662578 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a020010-9967-4da2-9f0e-6ec5fc461137" containerName="glance-log" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.663633 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.669715 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-2rt62" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.670155 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.670356 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.670545 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.679464 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.743010 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"527b1b7d-df6a-4d55-8aba-2f1333cd5534","Type":"ContainerDied","Data":"f0029dafde3d93ff6551c565442b5556f4dde57439b6c92dc5ad4146eef37c5c"} Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.743062 4787 scope.go:117] "RemoveContainer" containerID="0875d6598ced04ae5adf41e252c6f3988626ac39ce64ccae48a815ad87ee8791" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.743069 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.752014 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5c7b59dd4f-spxm4" event={"ID":"e743cd5e-e8c6-4fe2-9480-3a30316b8e23","Type":"ContainerStarted","Data":"9b76c0f3653d44a65a99f3347665c392b575fb40c81e4165c3c799ca7bb64398"} Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.764161 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7bd7989875-t7k69"] Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.765825 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bd7989875-t7k69" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.782592 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.783523 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c","Type":"ContainerDied","Data":"e3cc61e669df8493afec070a71fc990a8c3efae859954713f94e0f84e58643af"} Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.791350 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rkzz\" (UniqueName: \"kubernetes.io/projected/6eff6348-9415-4d61-a405-0a6ac27869d4-kube-api-access-8rkzz\") pod \"cinder-scheduler-0\" (UID: \"6eff6348-9415-4d61-a405-0a6ac27869d4\") " pod="openstack/cinder-scheduler-0" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.795486 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6eff6348-9415-4d61-a405-0a6ac27869d4-scripts\") pod \"cinder-scheduler-0\" (UID: \"6eff6348-9415-4d61-a405-0a6ac27869d4\") " pod="openstack/cinder-scheduler-0" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.795570 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff6348-9415-4d61-a405-0a6ac27869d4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6eff6348-9415-4d61-a405-0a6ac27869d4\") " pod="openstack/cinder-scheduler-0" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.795619 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6eff6348-9415-4d61-a405-0a6ac27869d4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6eff6348-9415-4d61-a405-0a6ac27869d4\") " pod="openstack/cinder-scheduler-0" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.795785 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6eff6348-9415-4d61-a405-0a6ac27869d4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6eff6348-9415-4d61-a405-0a6ac27869d4\") " pod="openstack/cinder-scheduler-0" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.795849 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6eff6348-9415-4d61-a405-0a6ac27869d4-config-data\") pod \"cinder-scheduler-0\" (UID: \"6eff6348-9415-4d61-a405-0a6ac27869d4\") " pod="openstack/cinder-scheduler-0" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.806985 4787 scope.go:117] "RemoveContainer" containerID="0fe7ee68c24d07552d2277601f96ad684aedcf7f34ffbb273dc8b5ea7a7d2da8" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.847533 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-nhmnw" event={"ID":"931c8c97-57b0-4b0c-a97d-498eaad48543","Type":"ContainerStarted","Data":"6c9d2abe0acc8461d7228c5f12c76da347f54d558bfe1892d9caa55030e4879a"} Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.847581 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-nhmnw" event={"ID":"931c8c97-57b0-4b0c-a97d-498eaad48543","Type":"ContainerStarted","Data":"b392312040e6c776e604a490758bc2c11eb31fe0e20e065a9609db1d7ed3dd1f"} Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.869403 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-69qw2" event={"ID":"66351c20-c4db-489c-857f-c82107d2be61","Type":"ContainerStarted","Data":"b9413fc51a4c3892f13d3a0c92848789cf1e05429822e02ff764a347dc393fdc"} Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.869477 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-69qw2" event={"ID":"66351c20-c4db-489c-857f-c82107d2be61","Type":"ContainerStarted","Data":"4003d2ccdbec7e26bf11a45fc4ed672d6eda38caa2c6c7727d04cf46d784a1c1"} Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.873644 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bd7989875-t7k69"] Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.876587 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1a020010-9967-4da2-9f0e-6ec5fc461137","Type":"ContainerDied","Data":"5079c9b6b1b9bcfaac45797acdb2dcebf8fc508abe65ba630e5336fd6866ddc9"} Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.876726 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.891162 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.897304 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6eff6348-9415-4d61-a405-0a6ac27869d4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6eff6348-9415-4d61-a405-0a6ac27869d4\") " pod="openstack/cinder-scheduler-0" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.897365 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/951b7125-da65-4cfa-ae70-bd8013441bbb-ovsdbserver-sb\") pod \"dnsmasq-dns-7bd7989875-t7k69\" (UID: \"951b7125-da65-4cfa-ae70-bd8013441bbb\") " pod="openstack/dnsmasq-dns-7bd7989875-t7k69" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.897391 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/951b7125-da65-4cfa-ae70-bd8013441bbb-dns-swift-storage-0\") pod \"dnsmasq-dns-7bd7989875-t7k69\" (UID: \"951b7125-da65-4cfa-ae70-bd8013441bbb\") " pod="openstack/dnsmasq-dns-7bd7989875-t7k69" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.897445 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6eff6348-9415-4d61-a405-0a6ac27869d4-config-data\") pod \"cinder-scheduler-0\" (UID: \"6eff6348-9415-4d61-a405-0a6ac27869d4\") " pod="openstack/cinder-scheduler-0" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.897476 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4bpb\" (UniqueName: \"kubernetes.io/projected/951b7125-da65-4cfa-ae70-bd8013441bbb-kube-api-access-q4bpb\") pod \"dnsmasq-dns-7bd7989875-t7k69\" (UID: \"951b7125-da65-4cfa-ae70-bd8013441bbb\") " pod="openstack/dnsmasq-dns-7bd7989875-t7k69" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.897508 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/951b7125-da65-4cfa-ae70-bd8013441bbb-dns-svc\") pod \"dnsmasq-dns-7bd7989875-t7k69\" (UID: \"951b7125-da65-4cfa-ae70-bd8013441bbb\") " pod="openstack/dnsmasq-dns-7bd7989875-t7k69" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.897525 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rkzz\" (UniqueName: \"kubernetes.io/projected/6eff6348-9415-4d61-a405-0a6ac27869d4-kube-api-access-8rkzz\") pod \"cinder-scheduler-0\" (UID: \"6eff6348-9415-4d61-a405-0a6ac27869d4\") " pod="openstack/cinder-scheduler-0" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.897556 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6eff6348-9415-4d61-a405-0a6ac27869d4-scripts\") pod \"cinder-scheduler-0\" (UID: \"6eff6348-9415-4d61-a405-0a6ac27869d4\") " pod="openstack/cinder-scheduler-0" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.897580 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff6348-9415-4d61-a405-0a6ac27869d4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6eff6348-9415-4d61-a405-0a6ac27869d4\") " pod="openstack/cinder-scheduler-0" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.897602 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6eff6348-9415-4d61-a405-0a6ac27869d4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6eff6348-9415-4d61-a405-0a6ac27869d4\") " pod="openstack/cinder-scheduler-0" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.897627 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/951b7125-da65-4cfa-ae70-bd8013441bbb-config\") pod \"dnsmasq-dns-7bd7989875-t7k69\" (UID: \"951b7125-da65-4cfa-ae70-bd8013441bbb\") " pod="openstack/dnsmasq-dns-7bd7989875-t7k69" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.897653 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/951b7125-da65-4cfa-ae70-bd8013441bbb-ovsdbserver-nb\") pod \"dnsmasq-dns-7bd7989875-t7k69\" (UID: \"951b7125-da65-4cfa-ae70-bd8013441bbb\") " pod="openstack/dnsmasq-dns-7bd7989875-t7k69" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.902151 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6eff6348-9415-4d61-a405-0a6ac27869d4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6eff6348-9415-4d61-a405-0a6ac27869d4\") " pod="openstack/cinder-scheduler-0" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.902734 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6eff6348-9415-4d61-a405-0a6ac27869d4-config-data\") pod \"cinder-scheduler-0\" (UID: \"6eff6348-9415-4d61-a405-0a6ac27869d4\") " pod="openstack/cinder-scheduler-0" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.907339 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6eff6348-9415-4d61-a405-0a6ac27869d4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6eff6348-9415-4d61-a405-0a6ac27869d4\") " pod="openstack/cinder-scheduler-0" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.908368 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"ccb620ed-ad28-4ec7-8319-6d5bb3ce10ec","Type":"ContainerStarted","Data":"b679fe7854243d66ad08c72cb71af278da09827d6095c4050471e1bdfce26dfa"} Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.922746 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff6348-9415-4d61-a405-0a6ac27869d4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6eff6348-9415-4d61-a405-0a6ac27869d4\") " pod="openstack/cinder-scheduler-0" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.923445 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6eff6348-9415-4d61-a405-0a6ac27869d4-scripts\") pod \"cinder-scheduler-0\" (UID: \"6eff6348-9415-4d61-a405-0a6ac27869d4\") " pod="openstack/cinder-scheduler-0" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.931455 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-9kp59" event={"ID":"a7ba3660-dca6-4843-b560-efd0ea158258","Type":"ContainerStarted","Data":"998de087f5256977e33199f2b4e1481765a110036bd7b68cb2f9de0a37109a9e"} Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.931517 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-9kp59" event={"ID":"a7ba3660-dca6-4843-b560-efd0ea158258","Type":"ContainerStarted","Data":"5660f5fe231782e3bf4856360e01d1364c9068560fee8615dcb01eb0270187f2"} Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.932106 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rkzz\" (UniqueName: \"kubernetes.io/projected/6eff6348-9415-4d61-a405-0a6ac27869d4-kube-api-access-8rkzz\") pod \"cinder-scheduler-0\" (UID: \"6eff6348-9415-4d61-a405-0a6ac27869d4\") " pod="openstack/cinder-scheduler-0" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.942451 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.955098 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-79b69df69b-mfcjt" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.955787 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.960967 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-79b69df69b-mfcjt" event={"ID":"0244e33a-702c-412d-8074-a2c910f375d0","Type":"ContainerDied","Data":"5aed0b8e30923bdfcedfb7b6f9d63cadbeed4c395541d46dd2ec706babcdfe89"} Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.961153 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.966835 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-bskck" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.967661 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.967710 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.967795 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.988011 4787 scope.go:117] "RemoveContainer" containerID="ae68e6f7fc578d6c34f91b087188337008568de65db874e1e8a7966190fdc84e" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.990759 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.999950 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/951b7125-da65-4cfa-ae70-bd8013441bbb-config\") pod \"dnsmasq-dns-7bd7989875-t7k69\" (UID: \"951b7125-da65-4cfa-ae70-bd8013441bbb\") " pod="openstack/dnsmasq-dns-7bd7989875-t7k69" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:22.999997 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/951b7125-da65-4cfa-ae70-bd8013441bbb-ovsdbserver-nb\") pod \"dnsmasq-dns-7bd7989875-t7k69\" (UID: \"951b7125-da65-4cfa-ae70-bd8013441bbb\") " pod="openstack/dnsmasq-dns-7bd7989875-t7k69" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:23.000088 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/951b7125-da65-4cfa-ae70-bd8013441bbb-ovsdbserver-sb\") pod \"dnsmasq-dns-7bd7989875-t7k69\" (UID: \"951b7125-da65-4cfa-ae70-bd8013441bbb\") " pod="openstack/dnsmasq-dns-7bd7989875-t7k69" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:23.000110 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/951b7125-da65-4cfa-ae70-bd8013441bbb-dns-swift-storage-0\") pod \"dnsmasq-dns-7bd7989875-t7k69\" (UID: \"951b7125-da65-4cfa-ae70-bd8013441bbb\") " pod="openstack/dnsmasq-dns-7bd7989875-t7k69" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:23.000142 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4bpb\" (UniqueName: \"kubernetes.io/projected/951b7125-da65-4cfa-ae70-bd8013441bbb-kube-api-access-q4bpb\") pod \"dnsmasq-dns-7bd7989875-t7k69\" (UID: \"951b7125-da65-4cfa-ae70-bd8013441bbb\") " pod="openstack/dnsmasq-dns-7bd7989875-t7k69" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:23.000188 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/951b7125-da65-4cfa-ae70-bd8013441bbb-dns-svc\") pod \"dnsmasq-dns-7bd7989875-t7k69\" (UID: \"951b7125-da65-4cfa-ae70-bd8013441bbb\") " pod="openstack/dnsmasq-dns-7bd7989875-t7k69" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:23.003519 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/951b7125-da65-4cfa-ae70-bd8013441bbb-dns-svc\") pod \"dnsmasq-dns-7bd7989875-t7k69\" (UID: \"951b7125-da65-4cfa-ae70-bd8013441bbb\") " pod="openstack/dnsmasq-dns-7bd7989875-t7k69" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:23.003823 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:23.003851 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/951b7125-da65-4cfa-ae70-bd8013441bbb-ovsdbserver-sb\") pod \"dnsmasq-dns-7bd7989875-t7k69\" (UID: \"951b7125-da65-4cfa-ae70-bd8013441bbb\") " pod="openstack/dnsmasq-dns-7bd7989875-t7k69" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:23.004092 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/951b7125-da65-4cfa-ae70-bd8013441bbb-dns-swift-storage-0\") pod \"dnsmasq-dns-7bd7989875-t7k69\" (UID: \"951b7125-da65-4cfa-ae70-bd8013441bbb\") " pod="openstack/dnsmasq-dns-7bd7989875-t7k69" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:23.004409 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/951b7125-da65-4cfa-ae70-bd8013441bbb-config\") pod \"dnsmasq-dns-7bd7989875-t7k69\" (UID: \"951b7125-da65-4cfa-ae70-bd8013441bbb\") " pod="openstack/dnsmasq-dns-7bd7989875-t7k69" Oct 01 09:53:22 crc kubenswrapper[4787]: I1001 09:53:23.004650 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/951b7125-da65-4cfa-ae70-bd8013441bbb-ovsdbserver-nb\") pod \"dnsmasq-dns-7bd7989875-t7k69\" (UID: \"951b7125-da65-4cfa-ae70-bd8013441bbb\") " pod="openstack/dnsmasq-dns-7bd7989875-t7k69" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.023998 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.025719 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.030759 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-nhmnw" podStartSLOduration=6.030725221 podStartE2EDuration="6.030725221s" podCreationTimestamp="2025-10-01 09:53:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:53:22.864248226 +0000 UTC m=+1034.979392403" watchObservedRunningTime="2025-10-01 09:53:23.030725221 +0000 UTC m=+1035.145869378" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.031045 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.031350 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.075025 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4bpb\" (UniqueName: \"kubernetes.io/projected/951b7125-da65-4cfa-ae70-bd8013441bbb-kube-api-access-q4bpb\") pod \"dnsmasq-dns-7bd7989875-t7k69\" (UID: \"951b7125-da65-4cfa-ae70-bd8013441bbb\") " pod="openstack/dnsmasq-dns-7bd7989875-t7k69" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.076964 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.094250 4787 scope.go:117] "RemoveContainer" containerID="9ea5042ebd4a4cfdce6dd4dd8b3f130111459a9b369f9c3d2aa1cb0f72d20ad0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.102516 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9c9r\" (UniqueName: \"kubernetes.io/projected/ddca8826-066e-45d5-9536-59ef0bfc8c61-kube-api-access-w9c9r\") pod \"glance-default-internal-api-0\" (UID: \"ddca8826-066e-45d5-9536-59ef0bfc8c61\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.103501 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/165b7740-cef4-45af-87d1-901b882b8dbb-config-data\") pod \"cinder-api-0\" (UID: \"165b7740-cef4-45af-87d1-901b882b8dbb\") " pod="openstack/cinder-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.103901 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"ddca8826-066e-45d5-9536-59ef0bfc8c61\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.103939 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddca8826-066e-45d5-9536-59ef0bfc8c61-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ddca8826-066e-45d5-9536-59ef0bfc8c61\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.104008 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/165b7740-cef4-45af-87d1-901b882b8dbb-config-data-custom\") pod \"cinder-api-0\" (UID: \"165b7740-cef4-45af-87d1-901b882b8dbb\") " pod="openstack/cinder-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.104043 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/165b7740-cef4-45af-87d1-901b882b8dbb-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"165b7740-cef4-45af-87d1-901b882b8dbb\") " pod="openstack/cinder-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.104098 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ddca8826-066e-45d5-9536-59ef0bfc8c61-logs\") pod \"glance-default-internal-api-0\" (UID: \"ddca8826-066e-45d5-9536-59ef0bfc8c61\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.104155 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/165b7740-cef4-45af-87d1-901b882b8dbb-logs\") pod \"cinder-api-0\" (UID: \"165b7740-cef4-45af-87d1-901b882b8dbb\") " pod="openstack/cinder-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.104219 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mg4lx\" (UniqueName: \"kubernetes.io/projected/165b7740-cef4-45af-87d1-901b882b8dbb-kube-api-access-mg4lx\") pod \"cinder-api-0\" (UID: \"165b7740-cef4-45af-87d1-901b882b8dbb\") " pod="openstack/cinder-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.104240 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddca8826-066e-45d5-9536-59ef0bfc8c61-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ddca8826-066e-45d5-9536-59ef0bfc8c61\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.104298 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/165b7740-cef4-45af-87d1-901b882b8dbb-scripts\") pod \"cinder-api-0\" (UID: \"165b7740-cef4-45af-87d1-901b882b8dbb\") " pod="openstack/cinder-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.104383 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddca8826-066e-45d5-9536-59ef0bfc8c61-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ddca8826-066e-45d5-9536-59ef0bfc8c61\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.104423 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/165b7740-cef4-45af-87d1-901b882b8dbb-etc-machine-id\") pod \"cinder-api-0\" (UID: \"165b7740-cef4-45af-87d1-901b882b8dbb\") " pod="openstack/cinder-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.104452 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ddca8826-066e-45d5-9536-59ef0bfc8c61-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ddca8826-066e-45d5-9536-59ef0bfc8c61\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.104508 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddca8826-066e-45d5-9536-59ef0bfc8c61-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ddca8826-066e-45d5-9536-59ef0bfc8c61\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.105673 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.122070 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.126031 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.129547 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.129728 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.150156 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.185453 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.493578324 podStartE2EDuration="16.185432279s" podCreationTimestamp="2025-10-01 09:53:07 +0000 UTC" firstStartedPulling="2025-10-01 09:53:08.745842989 +0000 UTC m=+1020.860987146" lastFinishedPulling="2025-10-01 09:53:21.437696784 +0000 UTC m=+1033.552841101" observedRunningTime="2025-10-01 09:53:22.983468264 +0000 UTC m=+1035.098612421" watchObservedRunningTime="2025-10-01 09:53:23.185432279 +0000 UTC m=+1035.300576436" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.186250 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-67879968b-d5598" podUID="7983e213-c317-42b2-8944-85bdb78dac8f" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.186383 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-67879968b-d5598" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.206991 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/165b7740-cef4-45af-87d1-901b882b8dbb-config-data-custom\") pod \"cinder-api-0\" (UID: \"165b7740-cef4-45af-87d1-901b882b8dbb\") " pod="openstack/cinder-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.207054 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/165b7740-cef4-45af-87d1-901b882b8dbb-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"165b7740-cef4-45af-87d1-901b882b8dbb\") " pod="openstack/cinder-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.207097 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-run-httpd\") pod \"ceilometer-0\" (UID: \"d117eceb-fcde-43be-b1be-d6acbfe6c0ac\") " pod="openstack/ceilometer-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.207134 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d117eceb-fcde-43be-b1be-d6acbfe6c0ac\") " pod="openstack/ceilometer-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.207169 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ddca8826-066e-45d5-9536-59ef0bfc8c61-logs\") pod \"glance-default-internal-api-0\" (UID: \"ddca8826-066e-45d5-9536-59ef0bfc8c61\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.207202 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/165b7740-cef4-45af-87d1-901b882b8dbb-logs\") pod \"cinder-api-0\" (UID: \"165b7740-cef4-45af-87d1-901b882b8dbb\") " pod="openstack/cinder-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.207236 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mg4lx\" (UniqueName: \"kubernetes.io/projected/165b7740-cef4-45af-87d1-901b882b8dbb-kube-api-access-mg4lx\") pod \"cinder-api-0\" (UID: \"165b7740-cef4-45af-87d1-901b882b8dbb\") " pod="openstack/cinder-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.207258 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddca8826-066e-45d5-9536-59ef0bfc8c61-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ddca8826-066e-45d5-9536-59ef0bfc8c61\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.207298 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/165b7740-cef4-45af-87d1-901b882b8dbb-scripts\") pod \"cinder-api-0\" (UID: \"165b7740-cef4-45af-87d1-901b882b8dbb\") " pod="openstack/cinder-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.207338 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-log-httpd\") pod \"ceilometer-0\" (UID: \"d117eceb-fcde-43be-b1be-d6acbfe6c0ac\") " pod="openstack/ceilometer-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.207371 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d117eceb-fcde-43be-b1be-d6acbfe6c0ac\") " pod="openstack/ceilometer-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.207398 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddca8826-066e-45d5-9536-59ef0bfc8c61-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ddca8826-066e-45d5-9536-59ef0bfc8c61\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.207521 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/165b7740-cef4-45af-87d1-901b882b8dbb-etc-machine-id\") pod \"cinder-api-0\" (UID: \"165b7740-cef4-45af-87d1-901b882b8dbb\") " pod="openstack/cinder-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.207552 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-scripts\") pod \"ceilometer-0\" (UID: \"d117eceb-fcde-43be-b1be-d6acbfe6c0ac\") " pod="openstack/ceilometer-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.207577 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ddca8826-066e-45d5-9536-59ef0bfc8c61-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ddca8826-066e-45d5-9536-59ef0bfc8c61\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.207613 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lw8dm\" (UniqueName: \"kubernetes.io/projected/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-kube-api-access-lw8dm\") pod \"ceilometer-0\" (UID: \"d117eceb-fcde-43be-b1be-d6acbfe6c0ac\") " pod="openstack/ceilometer-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.207648 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddca8826-066e-45d5-9536-59ef0bfc8c61-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ddca8826-066e-45d5-9536-59ef0bfc8c61\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.207677 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9c9r\" (UniqueName: \"kubernetes.io/projected/ddca8826-066e-45d5-9536-59ef0bfc8c61-kube-api-access-w9c9r\") pod \"glance-default-internal-api-0\" (UID: \"ddca8826-066e-45d5-9536-59ef0bfc8c61\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.207701 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/165b7740-cef4-45af-87d1-901b882b8dbb-config-data\") pod \"cinder-api-0\" (UID: \"165b7740-cef4-45af-87d1-901b882b8dbb\") " pod="openstack/cinder-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.207733 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-config-data\") pod \"ceilometer-0\" (UID: \"d117eceb-fcde-43be-b1be-d6acbfe6c0ac\") " pod="openstack/ceilometer-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.207776 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"ddca8826-066e-45d5-9536-59ef0bfc8c61\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.207805 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddca8826-066e-45d5-9536-59ef0bfc8c61-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ddca8826-066e-45d5-9536-59ef0bfc8c61\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.216300 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/165b7740-cef4-45af-87d1-901b882b8dbb-config-data-custom\") pod \"cinder-api-0\" (UID: \"165b7740-cef4-45af-87d1-901b882b8dbb\") " pod="openstack/cinder-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.227695 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/165b7740-cef4-45af-87d1-901b882b8dbb-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"165b7740-cef4-45af-87d1-901b882b8dbb\") " pod="openstack/cinder-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.238404 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ddca8826-066e-45d5-9536-59ef0bfc8c61-logs\") pod \"glance-default-internal-api-0\" (UID: \"ddca8826-066e-45d5-9536-59ef0bfc8c61\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.238678 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/165b7740-cef4-45af-87d1-901b882b8dbb-logs\") pod \"cinder-api-0\" (UID: \"165b7740-cef4-45af-87d1-901b882b8dbb\") " pod="openstack/cinder-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.241717 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"ddca8826-066e-45d5-9536-59ef0bfc8c61\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.243470 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/165b7740-cef4-45af-87d1-901b882b8dbb-etc-machine-id\") pod \"cinder-api-0\" (UID: \"165b7740-cef4-45af-87d1-901b882b8dbb\") " pod="openstack/cinder-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.243951 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ddca8826-066e-45d5-9536-59ef0bfc8c61-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ddca8826-066e-45d5-9536-59ef0bfc8c61\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.254484 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/165b7740-cef4-45af-87d1-901b882b8dbb-scripts\") pod \"cinder-api-0\" (UID: \"165b7740-cef4-45af-87d1-901b882b8dbb\") " pod="openstack/cinder-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.258358 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddca8826-066e-45d5-9536-59ef0bfc8c61-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ddca8826-066e-45d5-9536-59ef0bfc8c61\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.258954 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddca8826-066e-45d5-9536-59ef0bfc8c61-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ddca8826-066e-45d5-9536-59ef0bfc8c61\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.258954 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddca8826-066e-45d5-9536-59ef0bfc8c61-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ddca8826-066e-45d5-9536-59ef0bfc8c61\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.259801 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddca8826-066e-45d5-9536-59ef0bfc8c61-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ddca8826-066e-45d5-9536-59ef0bfc8c61\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.266539 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bd7989875-t7k69" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.266752 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mg4lx\" (UniqueName: \"kubernetes.io/projected/165b7740-cef4-45af-87d1-901b882b8dbb-kube-api-access-mg4lx\") pod \"cinder-api-0\" (UID: \"165b7740-cef4-45af-87d1-901b882b8dbb\") " pod="openstack/cinder-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.267657 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/165b7740-cef4-45af-87d1-901b882b8dbb-config-data\") pod \"cinder-api-0\" (UID: \"165b7740-cef4-45af-87d1-901b882b8dbb\") " pod="openstack/cinder-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.288279 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9c9r\" (UniqueName: \"kubernetes.io/projected/ddca8826-066e-45d5-9536-59ef0bfc8c61-kube-api-access-w9c9r\") pod \"glance-default-internal-api-0\" (UID: \"ddca8826-066e-45d5-9536-59ef0bfc8c61\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.311177 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-79b69df69b-mfcjt"] Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.315856 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-config-data\") pod \"ceilometer-0\" (UID: \"d117eceb-fcde-43be-b1be-d6acbfe6c0ac\") " pod="openstack/ceilometer-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.316057 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-run-httpd\") pod \"ceilometer-0\" (UID: \"d117eceb-fcde-43be-b1be-d6acbfe6c0ac\") " pod="openstack/ceilometer-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.316108 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d117eceb-fcde-43be-b1be-d6acbfe6c0ac\") " pod="openstack/ceilometer-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.316264 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-log-httpd\") pod \"ceilometer-0\" (UID: \"d117eceb-fcde-43be-b1be-d6acbfe6c0ac\") " pod="openstack/ceilometer-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.316326 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d117eceb-fcde-43be-b1be-d6acbfe6c0ac\") " pod="openstack/ceilometer-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.316393 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-scripts\") pod \"ceilometer-0\" (UID: \"d117eceb-fcde-43be-b1be-d6acbfe6c0ac\") " pod="openstack/ceilometer-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.316444 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lw8dm\" (UniqueName: \"kubernetes.io/projected/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-kube-api-access-lw8dm\") pod \"ceilometer-0\" (UID: \"d117eceb-fcde-43be-b1be-d6acbfe6c0ac\") " pod="openstack/ceilometer-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.324555 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-run-httpd\") pod \"ceilometer-0\" (UID: \"d117eceb-fcde-43be-b1be-d6acbfe6c0ac\") " pod="openstack/ceilometer-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.332898 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d117eceb-fcde-43be-b1be-d6acbfe6c0ac\") " pod="openstack/ceilometer-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.344847 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-log-httpd\") pod \"ceilometer-0\" (UID: \"d117eceb-fcde-43be-b1be-d6acbfe6c0ac\") " pod="openstack/ceilometer-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.350689 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-scripts\") pod \"ceilometer-0\" (UID: \"d117eceb-fcde-43be-b1be-d6acbfe6c0ac\") " pod="openstack/ceilometer-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.351396 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d117eceb-fcde-43be-b1be-d6acbfe6c0ac\") " pod="openstack/ceilometer-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.352257 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-config-data\") pod \"ceilometer-0\" (UID: \"d117eceb-fcde-43be-b1be-d6acbfe6c0ac\") " pod="openstack/ceilometer-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.369819 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lw8dm\" (UniqueName: \"kubernetes.io/projected/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-kube-api-access-lw8dm\") pod \"ceilometer-0\" (UID: \"d117eceb-fcde-43be-b1be-d6acbfe6c0ac\") " pod="openstack/ceilometer-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.376315 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.395999 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-79b69df69b-mfcjt"] Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.396827 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"ddca8826-066e-45d5-9536-59ef0bfc8c61\") " pod="openstack/glance-default-internal-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.429119 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.438890 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.440850 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.454711 4787 scope.go:117] "RemoveContainer" containerID="974cfa14c1d8f423c33e89b90de41d56b73acf24136c57d8454283e7000a8425" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.454845 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.540197 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.544809 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.547592 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.575805 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.593798 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.627472 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/137125b0-e356-4dbc-be53-2dbd3a06825f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"137125b0-e356-4dbc-be53-2dbd3a06825f\") " pod="openstack/glance-default-external-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.627563 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/137125b0-e356-4dbc-be53-2dbd3a06825f-logs\") pod \"glance-default-external-api-0\" (UID: \"137125b0-e356-4dbc-be53-2dbd3a06825f\") " pod="openstack/glance-default-external-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.627598 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/137125b0-e356-4dbc-be53-2dbd3a06825f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"137125b0-e356-4dbc-be53-2dbd3a06825f\") " pod="openstack/glance-default-external-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.627651 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j47qc\" (UniqueName: \"kubernetes.io/projected/137125b0-e356-4dbc-be53-2dbd3a06825f-kube-api-access-j47qc\") pod \"glance-default-external-api-0\" (UID: \"137125b0-e356-4dbc-be53-2dbd3a06825f\") " pod="openstack/glance-default-external-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.627694 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/137125b0-e356-4dbc-be53-2dbd3a06825f-scripts\") pod \"glance-default-external-api-0\" (UID: \"137125b0-e356-4dbc-be53-2dbd3a06825f\") " pod="openstack/glance-default-external-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.627714 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"137125b0-e356-4dbc-be53-2dbd3a06825f\") " pod="openstack/glance-default-external-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.627745 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/137125b0-e356-4dbc-be53-2dbd3a06825f-config-data\") pod \"glance-default-external-api-0\" (UID: \"137125b0-e356-4dbc-be53-2dbd3a06825f\") " pod="openstack/glance-default-external-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.627780 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/137125b0-e356-4dbc-be53-2dbd3a06825f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"137125b0-e356-4dbc-be53-2dbd3a06825f\") " pod="openstack/glance-default-external-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.648526 4787 scope.go:117] "RemoveContainer" containerID="98c33f6c04008cd30bdef34a1de003ea58ef86bd06cb44398a6e45d9e7916bc5" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.680691 4787 scope.go:117] "RemoveContainer" containerID="c1bdee63e5aab6c9222b16cd9ff592390efcad78109917777cf0033f2a45414c" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.729524 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/137125b0-e356-4dbc-be53-2dbd3a06825f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"137125b0-e356-4dbc-be53-2dbd3a06825f\") " pod="openstack/glance-default-external-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.729623 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/137125b0-e356-4dbc-be53-2dbd3a06825f-logs\") pod \"glance-default-external-api-0\" (UID: \"137125b0-e356-4dbc-be53-2dbd3a06825f\") " pod="openstack/glance-default-external-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.729691 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/137125b0-e356-4dbc-be53-2dbd3a06825f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"137125b0-e356-4dbc-be53-2dbd3a06825f\") " pod="openstack/glance-default-external-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.729784 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j47qc\" (UniqueName: \"kubernetes.io/projected/137125b0-e356-4dbc-be53-2dbd3a06825f-kube-api-access-j47qc\") pod \"glance-default-external-api-0\" (UID: \"137125b0-e356-4dbc-be53-2dbd3a06825f\") " pod="openstack/glance-default-external-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.729854 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/137125b0-e356-4dbc-be53-2dbd3a06825f-scripts\") pod \"glance-default-external-api-0\" (UID: \"137125b0-e356-4dbc-be53-2dbd3a06825f\") " pod="openstack/glance-default-external-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.729873 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"137125b0-e356-4dbc-be53-2dbd3a06825f\") " pod="openstack/glance-default-external-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.729899 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/137125b0-e356-4dbc-be53-2dbd3a06825f-config-data\") pod \"glance-default-external-api-0\" (UID: \"137125b0-e356-4dbc-be53-2dbd3a06825f\") " pod="openstack/glance-default-external-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.729926 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/137125b0-e356-4dbc-be53-2dbd3a06825f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"137125b0-e356-4dbc-be53-2dbd3a06825f\") " pod="openstack/glance-default-external-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.730523 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/137125b0-e356-4dbc-be53-2dbd3a06825f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"137125b0-e356-4dbc-be53-2dbd3a06825f\") " pod="openstack/glance-default-external-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.731366 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/137125b0-e356-4dbc-be53-2dbd3a06825f-logs\") pod \"glance-default-external-api-0\" (UID: \"137125b0-e356-4dbc-be53-2dbd3a06825f\") " pod="openstack/glance-default-external-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.731776 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"137125b0-e356-4dbc-be53-2dbd3a06825f\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.740531 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/137125b0-e356-4dbc-be53-2dbd3a06825f-config-data\") pod \"glance-default-external-api-0\" (UID: \"137125b0-e356-4dbc-be53-2dbd3a06825f\") " pod="openstack/glance-default-external-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.742442 4787 scope.go:117] "RemoveContainer" containerID="ddc800b2f9d11dcc930a8b39989f2bbc9f97a490a9ba0389ce5ceab59fe9641b" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.746806 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/137125b0-e356-4dbc-be53-2dbd3a06825f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"137125b0-e356-4dbc-be53-2dbd3a06825f\") " pod="openstack/glance-default-external-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.749828 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/137125b0-e356-4dbc-be53-2dbd3a06825f-scripts\") pod \"glance-default-external-api-0\" (UID: \"137125b0-e356-4dbc-be53-2dbd3a06825f\") " pod="openstack/glance-default-external-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.750377 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/137125b0-e356-4dbc-be53-2dbd3a06825f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"137125b0-e356-4dbc-be53-2dbd3a06825f\") " pod="openstack/glance-default-external-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.772854 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j47qc\" (UniqueName: \"kubernetes.io/projected/137125b0-e356-4dbc-be53-2dbd3a06825f-kube-api-access-j47qc\") pod \"glance-default-external-api-0\" (UID: \"137125b0-e356-4dbc-be53-2dbd3a06825f\") " pod="openstack/glance-default-external-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.790718 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"137125b0-e356-4dbc-be53-2dbd3a06825f\") " pod="openstack/glance-default-external-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.814225 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.817805 4787 scope.go:117] "RemoveContainer" containerID="9809e8583584726a8893b9ddf2134ade7d5b049f4935bc46a74fc053a250125d" Oct 01 09:53:23 crc kubenswrapper[4787]: W1001 09:53:23.856619 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6eff6348_9415_4d61_a405_0a6ac27869d4.slice/crio-1bd39e017189758cb0462d0ab61f7301e1c4322581543a1d3d63a9807fe40b93 WatchSource:0}: Error finding container 1bd39e017189758cb0462d0ab61f7301e1c4322581543a1d3d63a9807fe40b93: Status 404 returned error can't find the container with id 1bd39e017189758cb0462d0ab61f7301e1c4322581543a1d3d63a9807fe40b93 Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.897450 4787 scope.go:117] "RemoveContainer" containerID="f6e497aeeaf590b497d72aedd95d80d7aecc82ff93802c4e050cb6ed19e5e13c" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.930891 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.955447 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.978701 4787 generic.go:334] "Generic (PLEG): container finished" podID="931c8c97-57b0-4b0c-a97d-498eaad48543" containerID="6c9d2abe0acc8461d7228c5f12c76da347f54d558bfe1892d9caa55030e4879a" exitCode=0 Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.978891 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-nhmnw" event={"ID":"931c8c97-57b0-4b0c-a97d-498eaad48543","Type":"ContainerDied","Data":"6c9d2abe0acc8461d7228c5f12c76da347f54d558bfe1892d9caa55030e4879a"} Oct 01 09:53:23 crc kubenswrapper[4787]: W1001 09:53:23.983483 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod165b7740_cef4_45af_87d1_901b882b8dbb.slice/crio-f486490ee12a062a19ed9e615fce2bb3846504f077654e9f1430f0e9391ab71f WatchSource:0}: Error finding container f486490ee12a062a19ed9e615fce2bb3846504f077654e9f1430f0e9391ab71f: Status 404 returned error can't find the container with id f486490ee12a062a19ed9e615fce2bb3846504f077654e9f1430f0e9391ab71f Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.989265 4787 generic.go:334] "Generic (PLEG): container finished" podID="66351c20-c4db-489c-857f-c82107d2be61" containerID="b9413fc51a4c3892f13d3a0c92848789cf1e05429822e02ff764a347dc393fdc" exitCode=0 Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.989344 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-69qw2" event={"ID":"66351c20-c4db-489c-857f-c82107d2be61","Type":"ContainerDied","Data":"b9413fc51a4c3892f13d3a0c92848789cf1e05429822e02ff764a347dc393fdc"} Oct 01 09:53:23 crc kubenswrapper[4787]: I1001 09:53:23.992592 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6eff6348-9415-4d61-a405-0a6ac27869d4","Type":"ContainerStarted","Data":"1bd39e017189758cb0462d0ab61f7301e1c4322581543a1d3d63a9807fe40b93"} Oct 01 09:53:24 crc kubenswrapper[4787]: I1001 09:53:24.029026 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5c7b59dd4f-spxm4" event={"ID":"e743cd5e-e8c6-4fe2-9480-3a30316b8e23","Type":"ContainerStarted","Data":"1e98b02548dcb0eaf1af6392beb4c30f4c05fd0d3c99cf01b32c36082d1cd9b1"} Oct 01 09:53:24 crc kubenswrapper[4787]: I1001 09:53:24.029127 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5c7b59dd4f-spxm4" event={"ID":"e743cd5e-e8c6-4fe2-9480-3a30316b8e23","Type":"ContainerStarted","Data":"e311f50b9be8937bedc529904519584cabc83598ecf7fce664407b9aa3c57929"} Oct 01 09:53:24 crc kubenswrapper[4787]: I1001 09:53:24.030052 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5c7b59dd4f-spxm4" Oct 01 09:53:24 crc kubenswrapper[4787]: I1001 09:53:24.030114 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5c7b59dd4f-spxm4" Oct 01 09:53:24 crc kubenswrapper[4787]: I1001 09:53:24.047417 4787 generic.go:334] "Generic (PLEG): container finished" podID="a7ba3660-dca6-4843-b560-efd0ea158258" containerID="998de087f5256977e33199f2b4e1481765a110036bd7b68cb2f9de0a37109a9e" exitCode=0 Oct 01 09:53:24 crc kubenswrapper[4787]: I1001 09:53:24.047738 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-9kp59" event={"ID":"a7ba3660-dca6-4843-b560-efd0ea158258","Type":"ContainerDied","Data":"998de087f5256977e33199f2b4e1481765a110036bd7b68cb2f9de0a37109a9e"} Oct 01 09:53:24 crc kubenswrapper[4787]: I1001 09:53:24.050723 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-5c7b59dd4f-spxm4" podStartSLOduration=11.050702738 podStartE2EDuration="11.050702738s" podCreationTimestamp="2025-10-01 09:53:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:53:24.049926978 +0000 UTC m=+1036.165071135" watchObservedRunningTime="2025-10-01 09:53:24.050702738 +0000 UTC m=+1036.165846895" Oct 01 09:53:24 crc kubenswrapper[4787]: I1001 09:53:24.083489 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bd7989875-t7k69"] Oct 01 09:53:24 crc kubenswrapper[4787]: I1001 09:53:24.267017 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:53:24 crc kubenswrapper[4787]: I1001 09:53:24.556441 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-69qw2" Oct 01 09:53:24 crc kubenswrapper[4787]: I1001 09:53:24.572593 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0244e33a-702c-412d-8074-a2c910f375d0" path="/var/lib/kubelet/pods/0244e33a-702c-412d-8074-a2c910f375d0/volumes" Oct 01 09:53:24 crc kubenswrapper[4787]: I1001 09:53:24.573395 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a020010-9967-4da2-9f0e-6ec5fc461137" path="/var/lib/kubelet/pods/1a020010-9967-4da2-9f0e-6ec5fc461137/volumes" Oct 01 09:53:24 crc kubenswrapper[4787]: I1001 09:53:24.574304 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="527b1b7d-df6a-4d55-8aba-2f1333cd5534" path="/var/lib/kubelet/pods/527b1b7d-df6a-4d55-8aba-2f1333cd5534/volumes" Oct 01 09:53:24 crc kubenswrapper[4787]: I1001 09:53:24.575505 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c" path="/var/lib/kubelet/pods/b4ccc778-bb43-4e30-b9e2-01fe53ba7f3c/volumes" Oct 01 09:53:24 crc kubenswrapper[4787]: I1001 09:53:24.674058 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4sd2w\" (UniqueName: \"kubernetes.io/projected/66351c20-c4db-489c-857f-c82107d2be61-kube-api-access-4sd2w\") pod \"66351c20-c4db-489c-857f-c82107d2be61\" (UID: \"66351c20-c4db-489c-857f-c82107d2be61\") " Oct 01 09:53:24 crc kubenswrapper[4787]: I1001 09:53:24.687787 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66351c20-c4db-489c-857f-c82107d2be61-kube-api-access-4sd2w" (OuterVolumeSpecName: "kube-api-access-4sd2w") pod "66351c20-c4db-489c-857f-c82107d2be61" (UID: "66351c20-c4db-489c-857f-c82107d2be61"). InnerVolumeSpecName "kube-api-access-4sd2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:53:24 crc kubenswrapper[4787]: I1001 09:53:24.754473 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 01 09:53:24 crc kubenswrapper[4787]: I1001 09:53:24.782108 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4sd2w\" (UniqueName: \"kubernetes.io/projected/66351c20-c4db-489c-857f-c82107d2be61-kube-api-access-4sd2w\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:24 crc kubenswrapper[4787]: I1001 09:53:24.838256 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 01 09:53:24 crc kubenswrapper[4787]: W1001 09:53:24.840284 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podddca8826_066e_45d5_9536_59ef0bfc8c61.slice/crio-0e354a5a5db8183f09cc7ef73845e71d7b67c06d654dd082ddaac8f45e2fec71 WatchSource:0}: Error finding container 0e354a5a5db8183f09cc7ef73845e71d7b67c06d654dd082ddaac8f45e2fec71: Status 404 returned error can't find the container with id 0e354a5a5db8183f09cc7ef73845e71d7b67c06d654dd082ddaac8f45e2fec71 Oct 01 09:53:24 crc kubenswrapper[4787]: I1001 09:53:24.853551 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-9kp59" Oct 01 09:53:24 crc kubenswrapper[4787]: I1001 09:53:24.883094 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6kpmn\" (UniqueName: \"kubernetes.io/projected/a7ba3660-dca6-4843-b560-efd0ea158258-kube-api-access-6kpmn\") pod \"a7ba3660-dca6-4843-b560-efd0ea158258\" (UID: \"a7ba3660-dca6-4843-b560-efd0ea158258\") " Oct 01 09:53:24 crc kubenswrapper[4787]: I1001 09:53:24.887309 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7ba3660-dca6-4843-b560-efd0ea158258-kube-api-access-6kpmn" (OuterVolumeSpecName: "kube-api-access-6kpmn") pod "a7ba3660-dca6-4843-b560-efd0ea158258" (UID: "a7ba3660-dca6-4843-b560-efd0ea158258"). InnerVolumeSpecName "kube-api-access-6kpmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:53:24 crc kubenswrapper[4787]: I1001 09:53:24.986685 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6kpmn\" (UniqueName: \"kubernetes.io/projected/a7ba3660-dca6-4843-b560-efd0ea158258-kube-api-access-6kpmn\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:25 crc kubenswrapper[4787]: I1001 09:53:25.101488 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-9kp59" event={"ID":"a7ba3660-dca6-4843-b560-efd0ea158258","Type":"ContainerDied","Data":"5660f5fe231782e3bf4856360e01d1364c9068560fee8615dcb01eb0270187f2"} Oct 01 09:53:25 crc kubenswrapper[4787]: I1001 09:53:25.101531 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5660f5fe231782e3bf4856360e01d1364c9068560fee8615dcb01eb0270187f2" Oct 01 09:53:25 crc kubenswrapper[4787]: I1001 09:53:25.101593 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-9kp59" Oct 01 09:53:25 crc kubenswrapper[4787]: I1001 09:53:25.105826 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"165b7740-cef4-45af-87d1-901b882b8dbb","Type":"ContainerStarted","Data":"78550135629d670c8f5b64631408f1954e83c8f14e7e144995643c387b8f42dd"} Oct 01 09:53:25 crc kubenswrapper[4787]: I1001 09:53:25.105869 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"165b7740-cef4-45af-87d1-901b882b8dbb","Type":"ContainerStarted","Data":"f486490ee12a062a19ed9e615fce2bb3846504f077654e9f1430f0e9391ab71f"} Oct 01 09:53:25 crc kubenswrapper[4787]: I1001 09:53:25.107979 4787 generic.go:334] "Generic (PLEG): container finished" podID="951b7125-da65-4cfa-ae70-bd8013441bbb" containerID="b55726dfcd0b4cf35b279f0b055e1b5ce6a6a33e7784cc6dba6458a9b93b74b7" exitCode=0 Oct 01 09:53:25 crc kubenswrapper[4787]: I1001 09:53:25.108020 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bd7989875-t7k69" event={"ID":"951b7125-da65-4cfa-ae70-bd8013441bbb","Type":"ContainerDied","Data":"b55726dfcd0b4cf35b279f0b055e1b5ce6a6a33e7784cc6dba6458a9b93b74b7"} Oct 01 09:53:25 crc kubenswrapper[4787]: I1001 09:53:25.108035 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bd7989875-t7k69" event={"ID":"951b7125-da65-4cfa-ae70-bd8013441bbb","Type":"ContainerStarted","Data":"e2da0c0e33808d07d5aeec7061eac40ace897a5c666adde108505fad19e83eec"} Oct 01 09:53:25 crc kubenswrapper[4787]: I1001 09:53:25.163206 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ddca8826-066e-45d5-9536-59ef0bfc8c61","Type":"ContainerStarted","Data":"0e354a5a5db8183f09cc7ef73845e71d7b67c06d654dd082ddaac8f45e2fec71"} Oct 01 09:53:25 crc kubenswrapper[4787]: I1001 09:53:25.167058 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-69qw2" Oct 01 09:53:25 crc kubenswrapper[4787]: I1001 09:53:25.167559 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-69qw2" event={"ID":"66351c20-c4db-489c-857f-c82107d2be61","Type":"ContainerDied","Data":"4003d2ccdbec7e26bf11a45fc4ed672d6eda38caa2c6c7727d04cf46d784a1c1"} Oct 01 09:53:25 crc kubenswrapper[4787]: I1001 09:53:25.167784 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4003d2ccdbec7e26bf11a45fc4ed672d6eda38caa2c6c7727d04cf46d784a1c1" Oct 01 09:53:25 crc kubenswrapper[4787]: I1001 09:53:25.171672 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"137125b0-e356-4dbc-be53-2dbd3a06825f","Type":"ContainerStarted","Data":"29aa78d2780e67cdd14dedbd93c458a722f14cfaeb77722026abd62b3b7da505"} Oct 01 09:53:25 crc kubenswrapper[4787]: I1001 09:53:25.199176 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d117eceb-fcde-43be-b1be-d6acbfe6c0ac","Type":"ContainerStarted","Data":"4a202fc268274228b45839320fe265d7a6349eebc719049529e5e1a1cabf4e35"} Oct 01 09:53:25 crc kubenswrapper[4787]: I1001 09:53:25.430376 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 01 09:53:25 crc kubenswrapper[4787]: I1001 09:53:25.804366 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-nhmnw" Oct 01 09:53:25 crc kubenswrapper[4787]: I1001 09:53:25.928516 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vxnh\" (UniqueName: \"kubernetes.io/projected/931c8c97-57b0-4b0c-a97d-498eaad48543-kube-api-access-7vxnh\") pod \"931c8c97-57b0-4b0c-a97d-498eaad48543\" (UID: \"931c8c97-57b0-4b0c-a97d-498eaad48543\") " Oct 01 09:53:25 crc kubenswrapper[4787]: I1001 09:53:25.941213 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/931c8c97-57b0-4b0c-a97d-498eaad48543-kube-api-access-7vxnh" (OuterVolumeSpecName: "kube-api-access-7vxnh") pod "931c8c97-57b0-4b0c-a97d-498eaad48543" (UID: "931c8c97-57b0-4b0c-a97d-498eaad48543"). InnerVolumeSpecName "kube-api-access-7vxnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:53:26 crc kubenswrapper[4787]: I1001 09:53:26.031575 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vxnh\" (UniqueName: \"kubernetes.io/projected/931c8c97-57b0-4b0c-a97d-498eaad48543-kube-api-access-7vxnh\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:26 crc kubenswrapper[4787]: I1001 09:53:26.238506 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-nhmnw" event={"ID":"931c8c97-57b0-4b0c-a97d-498eaad48543","Type":"ContainerDied","Data":"b392312040e6c776e604a490758bc2c11eb31fe0e20e065a9609db1d7ed3dd1f"} Oct 01 09:53:26 crc kubenswrapper[4787]: I1001 09:53:26.238566 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b392312040e6c776e604a490758bc2c11eb31fe0e20e065a9609db1d7ed3dd1f" Oct 01 09:53:26 crc kubenswrapper[4787]: I1001 09:53:26.238525 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-nhmnw" Oct 01 09:53:26 crc kubenswrapper[4787]: I1001 09:53:26.245846 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d117eceb-fcde-43be-b1be-d6acbfe6c0ac","Type":"ContainerStarted","Data":"082bb82e79dc55abe18103279f1bf58f67b45b73bb7d96f7933aa0d852a259f5"} Oct 01 09:53:26 crc kubenswrapper[4787]: I1001 09:53:26.255902 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bd7989875-t7k69" event={"ID":"951b7125-da65-4cfa-ae70-bd8013441bbb","Type":"ContainerStarted","Data":"fabf0f56b0adabd51f151a10a9e41c01c0cfce1054514c925a914a256348f858"} Oct 01 09:53:26 crc kubenswrapper[4787]: I1001 09:53:26.256166 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7bd7989875-t7k69" Oct 01 09:53:26 crc kubenswrapper[4787]: I1001 09:53:26.280061 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7bd7989875-t7k69" podStartSLOduration=4.279979815 podStartE2EDuration="4.279979815s" podCreationTimestamp="2025-10-01 09:53:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:53:26.275668756 +0000 UTC m=+1038.390812923" watchObservedRunningTime="2025-10-01 09:53:26.279979815 +0000 UTC m=+1038.395123972" Oct 01 09:53:26 crc kubenswrapper[4787]: I1001 09:53:26.433053 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6b69d7b58f-t9lc6" podUID="6808314c-0e06-4fa4-9c3e-269e5952fd3e" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.154:5353: i/o timeout" Oct 01 09:53:27 crc kubenswrapper[4787]: I1001 09:53:27.283852 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d117eceb-fcde-43be-b1be-d6acbfe6c0ac","Type":"ContainerStarted","Data":"f8660db311600b14f532f6d7a168f19b043ddc07ac600d493be40f786ac58844"} Oct 01 09:53:27 crc kubenswrapper[4787]: I1001 09:53:27.292625 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"165b7740-cef4-45af-87d1-901b882b8dbb","Type":"ContainerStarted","Data":"5b8d571beb65892dfda779a98c015ddf058f85e29dff6824e5572a7889aa7046"} Oct 01 09:53:27 crc kubenswrapper[4787]: I1001 09:53:27.292803 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="165b7740-cef4-45af-87d1-901b882b8dbb" containerName="cinder-api-log" containerID="cri-o://78550135629d670c8f5b64631408f1954e83c8f14e7e144995643c387b8f42dd" gracePeriod=30 Oct 01 09:53:27 crc kubenswrapper[4787]: I1001 09:53:27.292899 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 01 09:53:27 crc kubenswrapper[4787]: I1001 09:53:27.293139 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="165b7740-cef4-45af-87d1-901b882b8dbb" containerName="cinder-api" containerID="cri-o://5b8d571beb65892dfda779a98c015ddf058f85e29dff6824e5572a7889aa7046" gracePeriod=30 Oct 01 09:53:27 crc kubenswrapper[4787]: I1001 09:53:27.329966 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.329949621 podStartE2EDuration="5.329949621s" podCreationTimestamp="2025-10-01 09:53:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:53:27.323119218 +0000 UTC m=+1039.438263385" watchObservedRunningTime="2025-10-01 09:53:27.329949621 +0000 UTC m=+1039.445093778" Oct 01 09:53:27 crc kubenswrapper[4787]: I1001 09:53:27.340560 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-858fdb94fd-6tpq6" Oct 01 09:53:27 crc kubenswrapper[4787]: I1001 09:53:27.350531 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ddca8826-066e-45d5-9536-59ef0bfc8c61","Type":"ContainerStarted","Data":"7c553ae9e15d936c8458279bffeba9be0f1da1c9de0eb6c5d188ed46323ae7df"} Oct 01 09:53:27 crc kubenswrapper[4787]: I1001 09:53:27.359688 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"137125b0-e356-4dbc-be53-2dbd3a06825f","Type":"ContainerStarted","Data":"9f95a7c4a6dc6fa9f8b2142b731a963d9993d28965210f2d184705d1c22f52bb"} Oct 01 09:53:27 crc kubenswrapper[4787]: I1001 09:53:27.402255 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6eff6348-9415-4d61-a405-0a6ac27869d4","Type":"ContainerStarted","Data":"5784ec8cb42dd5d8922a1fcfb5e4e0b56515623a7de4cb21e67610f768fd5bc3"} Oct 01 09:53:27 crc kubenswrapper[4787]: I1001 09:53:27.859209 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-400d-account-create-mtvjl"] Oct 01 09:53:27 crc kubenswrapper[4787]: E1001 09:53:27.860135 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="931c8c97-57b0-4b0c-a97d-498eaad48543" containerName="mariadb-database-create" Oct 01 09:53:27 crc kubenswrapper[4787]: I1001 09:53:27.860150 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="931c8c97-57b0-4b0c-a97d-498eaad48543" containerName="mariadb-database-create" Oct 01 09:53:27 crc kubenswrapper[4787]: E1001 09:53:27.860178 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7ba3660-dca6-4843-b560-efd0ea158258" containerName="mariadb-database-create" Oct 01 09:53:27 crc kubenswrapper[4787]: I1001 09:53:27.860185 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7ba3660-dca6-4843-b560-efd0ea158258" containerName="mariadb-database-create" Oct 01 09:53:27 crc kubenswrapper[4787]: E1001 09:53:27.860194 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66351c20-c4db-489c-857f-c82107d2be61" containerName="mariadb-database-create" Oct 01 09:53:27 crc kubenswrapper[4787]: I1001 09:53:27.860199 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="66351c20-c4db-489c-857f-c82107d2be61" containerName="mariadb-database-create" Oct 01 09:53:27 crc kubenswrapper[4787]: I1001 09:53:27.860357 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="66351c20-c4db-489c-857f-c82107d2be61" containerName="mariadb-database-create" Oct 01 09:53:27 crc kubenswrapper[4787]: I1001 09:53:27.860373 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="931c8c97-57b0-4b0c-a97d-498eaad48543" containerName="mariadb-database-create" Oct 01 09:53:27 crc kubenswrapper[4787]: I1001 09:53:27.860389 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7ba3660-dca6-4843-b560-efd0ea158258" containerName="mariadb-database-create" Oct 01 09:53:27 crc kubenswrapper[4787]: I1001 09:53:27.860963 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-400d-account-create-mtvjl" Oct 01 09:53:27 crc kubenswrapper[4787]: I1001 09:53:27.867437 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 01 09:53:27 crc kubenswrapper[4787]: I1001 09:53:27.884238 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-400d-account-create-mtvjl"] Oct 01 09:53:27 crc kubenswrapper[4787]: I1001 09:53:27.986706 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wl7qz\" (UniqueName: \"kubernetes.io/projected/0d768476-fd86-440c-9c5d-dd62b6cbcaa5-kube-api-access-wl7qz\") pod \"nova-api-400d-account-create-mtvjl\" (UID: \"0d768476-fd86-440c-9c5d-dd62b6cbcaa5\") " pod="openstack/nova-api-400d-account-create-mtvjl" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.076850 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67879968b-d5598" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.099887 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wl7qz\" (UniqueName: \"kubernetes.io/projected/0d768476-fd86-440c-9c5d-dd62b6cbcaa5-kube-api-access-wl7qz\") pod \"nova-api-400d-account-create-mtvjl\" (UID: \"0d768476-fd86-440c-9c5d-dd62b6cbcaa5\") " pod="openstack/nova-api-400d-account-create-mtvjl" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.138224 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wl7qz\" (UniqueName: \"kubernetes.io/projected/0d768476-fd86-440c-9c5d-dd62b6cbcaa5-kube-api-access-wl7qz\") pod \"nova-api-400d-account-create-mtvjl\" (UID: \"0d768476-fd86-440c-9c5d-dd62b6cbcaa5\") " pod="openstack/nova-api-400d-account-create-mtvjl" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.202502 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7983e213-c317-42b2-8944-85bdb78dac8f-logs\") pod \"7983e213-c317-42b2-8944-85bdb78dac8f\" (UID: \"7983e213-c317-42b2-8944-85bdb78dac8f\") " Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.203109 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7983e213-c317-42b2-8944-85bdb78dac8f-logs" (OuterVolumeSpecName: "logs") pod "7983e213-c317-42b2-8944-85bdb78dac8f" (UID: "7983e213-c317-42b2-8944-85bdb78dac8f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.233377 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7983e213-c317-42b2-8944-85bdb78dac8f-config-data\") pod \"7983e213-c317-42b2-8944-85bdb78dac8f\" (UID: \"7983e213-c317-42b2-8944-85bdb78dac8f\") " Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.230170 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-400d-account-create-mtvjl" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.233441 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7983e213-c317-42b2-8944-85bdb78dac8f-scripts\") pod \"7983e213-c317-42b2-8944-85bdb78dac8f\" (UID: \"7983e213-c317-42b2-8944-85bdb78dac8f\") " Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.233563 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7983e213-c317-42b2-8944-85bdb78dac8f-combined-ca-bundle\") pod \"7983e213-c317-42b2-8944-85bdb78dac8f\" (UID: \"7983e213-c317-42b2-8944-85bdb78dac8f\") " Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.233651 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gxpk\" (UniqueName: \"kubernetes.io/projected/7983e213-c317-42b2-8944-85bdb78dac8f-kube-api-access-5gxpk\") pod \"7983e213-c317-42b2-8944-85bdb78dac8f\" (UID: \"7983e213-c317-42b2-8944-85bdb78dac8f\") " Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.233719 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7983e213-c317-42b2-8944-85bdb78dac8f-horizon-secret-key\") pod \"7983e213-c317-42b2-8944-85bdb78dac8f\" (UID: \"7983e213-c317-42b2-8944-85bdb78dac8f\") " Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.233775 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/7983e213-c317-42b2-8944-85bdb78dac8f-horizon-tls-certs\") pod \"7983e213-c317-42b2-8944-85bdb78dac8f\" (UID: \"7983e213-c317-42b2-8944-85bdb78dac8f\") " Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.236214 4787 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7983e213-c317-42b2-8944-85bdb78dac8f-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.259281 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7983e213-c317-42b2-8944-85bdb78dac8f-kube-api-access-5gxpk" (OuterVolumeSpecName: "kube-api-access-5gxpk") pod "7983e213-c317-42b2-8944-85bdb78dac8f" (UID: "7983e213-c317-42b2-8944-85bdb78dac8f"). InnerVolumeSpecName "kube-api-access-5gxpk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.283315 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7983e213-c317-42b2-8944-85bdb78dac8f-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "7983e213-c317-42b2-8944-85bdb78dac8f" (UID: "7983e213-c317-42b2-8944-85bdb78dac8f"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.286925 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7983e213-c317-42b2-8944-85bdb78dac8f-config-data" (OuterVolumeSpecName: "config-data") pod "7983e213-c317-42b2-8944-85bdb78dac8f" (UID: "7983e213-c317-42b2-8944-85bdb78dac8f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.297503 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7983e213-c317-42b2-8944-85bdb78dac8f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7983e213-c317-42b2-8944-85bdb78dac8f" (UID: "7983e213-c317-42b2-8944-85bdb78dac8f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.333321 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7983e213-c317-42b2-8944-85bdb78dac8f-scripts" (OuterVolumeSpecName: "scripts") pod "7983e213-c317-42b2-8944-85bdb78dac8f" (UID: "7983e213-c317-42b2-8944-85bdb78dac8f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.337579 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7983e213-c317-42b2-8944-85bdb78dac8f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.337613 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gxpk\" (UniqueName: \"kubernetes.io/projected/7983e213-c317-42b2-8944-85bdb78dac8f-kube-api-access-5gxpk\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.337627 4787 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7983e213-c317-42b2-8944-85bdb78dac8f-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.337664 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7983e213-c317-42b2-8944-85bdb78dac8f-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.337674 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7983e213-c317-42b2-8944-85bdb78dac8f-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.380568 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7983e213-c317-42b2-8944-85bdb78dac8f-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "7983e213-c317-42b2-8944-85bdb78dac8f" (UID: "7983e213-c317-42b2-8944-85bdb78dac8f"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.393669 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5c7b59dd4f-spxm4" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.394337 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5c7b59dd4f-spxm4" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.399598 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.441274 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/165b7740-cef4-45af-87d1-901b882b8dbb-config-data-custom\") pod \"165b7740-cef4-45af-87d1-901b882b8dbb\" (UID: \"165b7740-cef4-45af-87d1-901b882b8dbb\") " Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.441348 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg4lx\" (UniqueName: \"kubernetes.io/projected/165b7740-cef4-45af-87d1-901b882b8dbb-kube-api-access-mg4lx\") pod \"165b7740-cef4-45af-87d1-901b882b8dbb\" (UID: \"165b7740-cef4-45af-87d1-901b882b8dbb\") " Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.441426 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/165b7740-cef4-45af-87d1-901b882b8dbb-etc-machine-id\") pod \"165b7740-cef4-45af-87d1-901b882b8dbb\" (UID: \"165b7740-cef4-45af-87d1-901b882b8dbb\") " Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.441457 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/165b7740-cef4-45af-87d1-901b882b8dbb-logs\") pod \"165b7740-cef4-45af-87d1-901b882b8dbb\" (UID: \"165b7740-cef4-45af-87d1-901b882b8dbb\") " Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.441576 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/165b7740-cef4-45af-87d1-901b882b8dbb-combined-ca-bundle\") pod \"165b7740-cef4-45af-87d1-901b882b8dbb\" (UID: \"165b7740-cef4-45af-87d1-901b882b8dbb\") " Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.441790 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/165b7740-cef4-45af-87d1-901b882b8dbb-scripts\") pod \"165b7740-cef4-45af-87d1-901b882b8dbb\" (UID: \"165b7740-cef4-45af-87d1-901b882b8dbb\") " Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.441852 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/165b7740-cef4-45af-87d1-901b882b8dbb-config-data\") pod \"165b7740-cef4-45af-87d1-901b882b8dbb\" (UID: \"165b7740-cef4-45af-87d1-901b882b8dbb\") " Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.442480 4787 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/7983e213-c317-42b2-8944-85bdb78dac8f-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.444934 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/165b7740-cef4-45af-87d1-901b882b8dbb-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "165b7740-cef4-45af-87d1-901b882b8dbb" (UID: "165b7740-cef4-45af-87d1-901b882b8dbb"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.447625 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/165b7740-cef4-45af-87d1-901b882b8dbb-logs" (OuterVolumeSpecName: "logs") pod "165b7740-cef4-45af-87d1-901b882b8dbb" (UID: "165b7740-cef4-45af-87d1-901b882b8dbb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.452403 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/165b7740-cef4-45af-87d1-901b882b8dbb-kube-api-access-mg4lx" (OuterVolumeSpecName: "kube-api-access-mg4lx") pod "165b7740-cef4-45af-87d1-901b882b8dbb" (UID: "165b7740-cef4-45af-87d1-901b882b8dbb"). InnerVolumeSpecName "kube-api-access-mg4lx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.453978 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/165b7740-cef4-45af-87d1-901b882b8dbb-scripts" (OuterVolumeSpecName: "scripts") pod "165b7740-cef4-45af-87d1-901b882b8dbb" (UID: "165b7740-cef4-45af-87d1-901b882b8dbb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.456794 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/165b7740-cef4-45af-87d1-901b882b8dbb-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "165b7740-cef4-45af-87d1-901b882b8dbb" (UID: "165b7740-cef4-45af-87d1-901b882b8dbb"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.466858 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ddca8826-066e-45d5-9536-59ef0bfc8c61","Type":"ContainerStarted","Data":"9ef1f215d1f74247b608a4d52aa4107f43c1c7af6f64b406747c9e734b6938a9"} Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.492198 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"137125b0-e356-4dbc-be53-2dbd3a06825f","Type":"ContainerStarted","Data":"1ae07bada7028af5faa31162918c13d3fe6c0980d54222a760c2979ed842a64e"} Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.499212 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6eff6348-9415-4d61-a405-0a6ac27869d4","Type":"ContainerStarted","Data":"c9ecc450edb50582561268ad6d9c1a2bd4afd58b50ac192de75e9b62a534e973"} Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.523169 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d117eceb-fcde-43be-b1be-d6acbfe6c0ac","Type":"ContainerStarted","Data":"831afab4a31e797b058e98f7db738f32e779ff985867fb2ce615b1d9f6bbe02a"} Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.529316 4787 generic.go:334] "Generic (PLEG): container finished" podID="7983e213-c317-42b2-8944-85bdb78dac8f" containerID="37b1a9eaf9bc59c2abf28488f9ac98956d1e78ecc9bdf3f5d70a1af066d4e549" exitCode=137 Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.529445 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67879968b-d5598" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.560728 4787 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/165b7740-cef4-45af-87d1-901b882b8dbb-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.560764 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg4lx\" (UniqueName: \"kubernetes.io/projected/165b7740-cef4-45af-87d1-901b882b8dbb-kube-api-access-mg4lx\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.560775 4787 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/165b7740-cef4-45af-87d1-901b882b8dbb-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.560819 4787 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/165b7740-cef4-45af-87d1-901b882b8dbb-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.560828 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/165b7740-cef4-45af-87d1-901b882b8dbb-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.565778 4787 generic.go:334] "Generic (PLEG): container finished" podID="165b7740-cef4-45af-87d1-901b882b8dbb" containerID="5b8d571beb65892dfda779a98c015ddf058f85e29dff6824e5572a7889aa7046" exitCode=0 Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.565811 4787 generic.go:334] "Generic (PLEG): container finished" podID="165b7740-cef4-45af-87d1-901b882b8dbb" containerID="78550135629d670c8f5b64631408f1954e83c8f14e7e144995643c387b8f42dd" exitCode=143 Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.566833 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.580417 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67879968b-d5598" event={"ID":"7983e213-c317-42b2-8944-85bdb78dac8f","Type":"ContainerDied","Data":"37b1a9eaf9bc59c2abf28488f9ac98956d1e78ecc9bdf3f5d70a1af066d4e549"} Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.580483 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67879968b-d5598" event={"ID":"7983e213-c317-42b2-8944-85bdb78dac8f","Type":"ContainerDied","Data":"b68c4125c2c2e1d0824ffc24ec8de859b0016496a3588a9f02f3ba487eb043ad"} Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.580495 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"165b7740-cef4-45af-87d1-901b882b8dbb","Type":"ContainerDied","Data":"5b8d571beb65892dfda779a98c015ddf058f85e29dff6824e5572a7889aa7046"} Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.580510 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"165b7740-cef4-45af-87d1-901b882b8dbb","Type":"ContainerDied","Data":"78550135629d670c8f5b64631408f1954e83c8f14e7e144995643c387b8f42dd"} Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.580519 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"165b7740-cef4-45af-87d1-901b882b8dbb","Type":"ContainerDied","Data":"f486490ee12a062a19ed9e615fce2bb3846504f077654e9f1430f0e9391ab71f"} Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.580553 4787 scope.go:117] "RemoveContainer" containerID="c363923f03c8431731ce61fb7088f6a51538ab7d45b2f5fb59d3d2d245ae02f0" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.581248 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/165b7740-cef4-45af-87d1-901b882b8dbb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "165b7740-cef4-45af-87d1-901b882b8dbb" (UID: "165b7740-cef4-45af-87d1-901b882b8dbb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.599225 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/165b7740-cef4-45af-87d1-901b882b8dbb-config-data" (OuterVolumeSpecName: "config-data") pod "165b7740-cef4-45af-87d1-901b882b8dbb" (UID: "165b7740-cef4-45af-87d1-901b882b8dbb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.643311 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.643288195 podStartE2EDuration="5.643288195s" podCreationTimestamp="2025-10-01 09:53:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:53:28.540062541 +0000 UTC m=+1040.655206698" watchObservedRunningTime="2025-10-01 09:53:28.643288195 +0000 UTC m=+1040.758432352" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.663839 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.6636712209999995 podStartE2EDuration="6.663671221s" podCreationTimestamp="2025-10-01 09:53:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:53:28.562721015 +0000 UTC m=+1040.677865202" watchObservedRunningTime="2025-10-01 09:53:28.663671221 +0000 UTC m=+1040.778815378" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.674744 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.675615843 podStartE2EDuration="6.674723851s" podCreationTimestamp="2025-10-01 09:53:22 +0000 UTC" firstStartedPulling="2025-10-01 09:53:23.916752076 +0000 UTC m=+1036.031896233" lastFinishedPulling="2025-10-01 09:53:24.915860094 +0000 UTC m=+1037.031004241" observedRunningTime="2025-10-01 09:53:28.587873402 +0000 UTC m=+1040.703017579" watchObservedRunningTime="2025-10-01 09:53:28.674723851 +0000 UTC m=+1040.789868008" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.689218 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/165b7740-cef4-45af-87d1-901b882b8dbb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.691725 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/165b7740-cef4-45af-87d1-901b882b8dbb-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.704445 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-67879968b-d5598"] Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.722252 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-67879968b-d5598"] Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.785540 4787 scope.go:117] "RemoveContainer" containerID="37b1a9eaf9bc59c2abf28488f9ac98956d1e78ecc9bdf3f5d70a1af066d4e549" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.840057 4787 scope.go:117] "RemoveContainer" containerID="c363923f03c8431731ce61fb7088f6a51538ab7d45b2f5fb59d3d2d245ae02f0" Oct 01 09:53:28 crc kubenswrapper[4787]: E1001 09:53:28.845210 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c363923f03c8431731ce61fb7088f6a51538ab7d45b2f5fb59d3d2d245ae02f0\": container with ID starting with c363923f03c8431731ce61fb7088f6a51538ab7d45b2f5fb59d3d2d245ae02f0 not found: ID does not exist" containerID="c363923f03c8431731ce61fb7088f6a51538ab7d45b2f5fb59d3d2d245ae02f0" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.845268 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c363923f03c8431731ce61fb7088f6a51538ab7d45b2f5fb59d3d2d245ae02f0"} err="failed to get container status \"c363923f03c8431731ce61fb7088f6a51538ab7d45b2f5fb59d3d2d245ae02f0\": rpc error: code = NotFound desc = could not find container \"c363923f03c8431731ce61fb7088f6a51538ab7d45b2f5fb59d3d2d245ae02f0\": container with ID starting with c363923f03c8431731ce61fb7088f6a51538ab7d45b2f5fb59d3d2d245ae02f0 not found: ID does not exist" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.845320 4787 scope.go:117] "RemoveContainer" containerID="37b1a9eaf9bc59c2abf28488f9ac98956d1e78ecc9bdf3f5d70a1af066d4e549" Oct 01 09:53:28 crc kubenswrapper[4787]: E1001 09:53:28.845703 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37b1a9eaf9bc59c2abf28488f9ac98956d1e78ecc9bdf3f5d70a1af066d4e549\": container with ID starting with 37b1a9eaf9bc59c2abf28488f9ac98956d1e78ecc9bdf3f5d70a1af066d4e549 not found: ID does not exist" containerID="37b1a9eaf9bc59c2abf28488f9ac98956d1e78ecc9bdf3f5d70a1af066d4e549" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.845727 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37b1a9eaf9bc59c2abf28488f9ac98956d1e78ecc9bdf3f5d70a1af066d4e549"} err="failed to get container status \"37b1a9eaf9bc59c2abf28488f9ac98956d1e78ecc9bdf3f5d70a1af066d4e549\": rpc error: code = NotFound desc = could not find container \"37b1a9eaf9bc59c2abf28488f9ac98956d1e78ecc9bdf3f5d70a1af066d4e549\": container with ID starting with 37b1a9eaf9bc59c2abf28488f9ac98956d1e78ecc9bdf3f5d70a1af066d4e549 not found: ID does not exist" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.845742 4787 scope.go:117] "RemoveContainer" containerID="5b8d571beb65892dfda779a98c015ddf058f85e29dff6824e5572a7889aa7046" Oct 01 09:53:28 crc kubenswrapper[4787]: W1001 09:53:28.847574 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d768476_fd86_440c_9c5d_dd62b6cbcaa5.slice/crio-56c3daa8184cadd6c5c67fc2582acd83288070a3c247066f29532439e5390de5 WatchSource:0}: Error finding container 56c3daa8184cadd6c5c67fc2582acd83288070a3c247066f29532439e5390de5: Status 404 returned error can't find the container with id 56c3daa8184cadd6c5c67fc2582acd83288070a3c247066f29532439e5390de5 Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.852413 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-400d-account-create-mtvjl"] Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.878650 4787 scope.go:117] "RemoveContainer" containerID="78550135629d670c8f5b64631408f1954e83c8f14e7e144995643c387b8f42dd" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.906484 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.921270 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.938480 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 01 09:53:28 crc kubenswrapper[4787]: E1001 09:53:28.939300 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="165b7740-cef4-45af-87d1-901b882b8dbb" containerName="cinder-api" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.939326 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="165b7740-cef4-45af-87d1-901b882b8dbb" containerName="cinder-api" Oct 01 09:53:28 crc kubenswrapper[4787]: E1001 09:53:28.939360 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7983e213-c317-42b2-8944-85bdb78dac8f" containerName="horizon" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.939367 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="7983e213-c317-42b2-8944-85bdb78dac8f" containerName="horizon" Oct 01 09:53:28 crc kubenswrapper[4787]: E1001 09:53:28.939379 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7983e213-c317-42b2-8944-85bdb78dac8f" containerName="horizon-log" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.939385 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="7983e213-c317-42b2-8944-85bdb78dac8f" containerName="horizon-log" Oct 01 09:53:28 crc kubenswrapper[4787]: E1001 09:53:28.939405 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="165b7740-cef4-45af-87d1-901b882b8dbb" containerName="cinder-api-log" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.939429 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="165b7740-cef4-45af-87d1-901b882b8dbb" containerName="cinder-api-log" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.939722 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="7983e213-c317-42b2-8944-85bdb78dac8f" containerName="horizon" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.939759 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="165b7740-cef4-45af-87d1-901b882b8dbb" containerName="cinder-api" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.939777 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="7983e213-c317-42b2-8944-85bdb78dac8f" containerName="horizon-log" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.939789 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="165b7740-cef4-45af-87d1-901b882b8dbb" containerName="cinder-api-log" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.941825 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.948063 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.951686 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.951742 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.951913 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.968756 4787 scope.go:117] "RemoveContainer" containerID="5b8d571beb65892dfda779a98c015ddf058f85e29dff6824e5572a7889aa7046" Oct 01 09:53:28 crc kubenswrapper[4787]: E1001 09:53:28.970156 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b8d571beb65892dfda779a98c015ddf058f85e29dff6824e5572a7889aa7046\": container with ID starting with 5b8d571beb65892dfda779a98c015ddf058f85e29dff6824e5572a7889aa7046 not found: ID does not exist" containerID="5b8d571beb65892dfda779a98c015ddf058f85e29dff6824e5572a7889aa7046" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.970196 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b8d571beb65892dfda779a98c015ddf058f85e29dff6824e5572a7889aa7046"} err="failed to get container status \"5b8d571beb65892dfda779a98c015ddf058f85e29dff6824e5572a7889aa7046\": rpc error: code = NotFound desc = could not find container \"5b8d571beb65892dfda779a98c015ddf058f85e29dff6824e5572a7889aa7046\": container with ID starting with 5b8d571beb65892dfda779a98c015ddf058f85e29dff6824e5572a7889aa7046 not found: ID does not exist" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.970226 4787 scope.go:117] "RemoveContainer" containerID="78550135629d670c8f5b64631408f1954e83c8f14e7e144995643c387b8f42dd" Oct 01 09:53:28 crc kubenswrapper[4787]: E1001 09:53:28.971246 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78550135629d670c8f5b64631408f1954e83c8f14e7e144995643c387b8f42dd\": container with ID starting with 78550135629d670c8f5b64631408f1954e83c8f14e7e144995643c387b8f42dd not found: ID does not exist" containerID="78550135629d670c8f5b64631408f1954e83c8f14e7e144995643c387b8f42dd" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.971270 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78550135629d670c8f5b64631408f1954e83c8f14e7e144995643c387b8f42dd"} err="failed to get container status \"78550135629d670c8f5b64631408f1954e83c8f14e7e144995643c387b8f42dd\": rpc error: code = NotFound desc = could not find container \"78550135629d670c8f5b64631408f1954e83c8f14e7e144995643c387b8f42dd\": container with ID starting with 78550135629d670c8f5b64631408f1954e83c8f14e7e144995643c387b8f42dd not found: ID does not exist" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.971284 4787 scope.go:117] "RemoveContainer" containerID="5b8d571beb65892dfda779a98c015ddf058f85e29dff6824e5572a7889aa7046" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.971452 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b8d571beb65892dfda779a98c015ddf058f85e29dff6824e5572a7889aa7046"} err="failed to get container status \"5b8d571beb65892dfda779a98c015ddf058f85e29dff6824e5572a7889aa7046\": rpc error: code = NotFound desc = could not find container \"5b8d571beb65892dfda779a98c015ddf058f85e29dff6824e5572a7889aa7046\": container with ID starting with 5b8d571beb65892dfda779a98c015ddf058f85e29dff6824e5572a7889aa7046 not found: ID does not exist" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.971474 4787 scope.go:117] "RemoveContainer" containerID="78550135629d670c8f5b64631408f1954e83c8f14e7e144995643c387b8f42dd" Oct 01 09:53:28 crc kubenswrapper[4787]: I1001 09:53:28.971655 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78550135629d670c8f5b64631408f1954e83c8f14e7e144995643c387b8f42dd"} err="failed to get container status \"78550135629d670c8f5b64631408f1954e83c8f14e7e144995643c387b8f42dd\": rpc error: code = NotFound desc = could not find container \"78550135629d670c8f5b64631408f1954e83c8f14e7e144995643c387b8f42dd\": container with ID starting with 78550135629d670c8f5b64631408f1954e83c8f14e7e144995643c387b8f42dd not found: ID does not exist" Oct 01 09:53:29 crc kubenswrapper[4787]: I1001 09:53:29.000662 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a05bd18f-4272-4ff1-ac85-65f0a20cbdcf-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"a05bd18f-4272-4ff1-ac85-65f0a20cbdcf\") " pod="openstack/cinder-api-0" Oct 01 09:53:29 crc kubenswrapper[4787]: I1001 09:53:29.001146 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a05bd18f-4272-4ff1-ac85-65f0a20cbdcf-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a05bd18f-4272-4ff1-ac85-65f0a20cbdcf\") " pod="openstack/cinder-api-0" Oct 01 09:53:29 crc kubenswrapper[4787]: I1001 09:53:29.001963 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a05bd18f-4272-4ff1-ac85-65f0a20cbdcf-public-tls-certs\") pod \"cinder-api-0\" (UID: \"a05bd18f-4272-4ff1-ac85-65f0a20cbdcf\") " pod="openstack/cinder-api-0" Oct 01 09:53:29 crc kubenswrapper[4787]: I1001 09:53:29.002031 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a05bd18f-4272-4ff1-ac85-65f0a20cbdcf-config-data\") pod \"cinder-api-0\" (UID: \"a05bd18f-4272-4ff1-ac85-65f0a20cbdcf\") " pod="openstack/cinder-api-0" Oct 01 09:53:29 crc kubenswrapper[4787]: I1001 09:53:29.002185 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a05bd18f-4272-4ff1-ac85-65f0a20cbdcf-logs\") pod \"cinder-api-0\" (UID: \"a05bd18f-4272-4ff1-ac85-65f0a20cbdcf\") " pod="openstack/cinder-api-0" Oct 01 09:53:29 crc kubenswrapper[4787]: I1001 09:53:29.002290 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a05bd18f-4272-4ff1-ac85-65f0a20cbdcf-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a05bd18f-4272-4ff1-ac85-65f0a20cbdcf\") " pod="openstack/cinder-api-0" Oct 01 09:53:29 crc kubenswrapper[4787]: I1001 09:53:29.002333 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jg5kt\" (UniqueName: \"kubernetes.io/projected/a05bd18f-4272-4ff1-ac85-65f0a20cbdcf-kube-api-access-jg5kt\") pod \"cinder-api-0\" (UID: \"a05bd18f-4272-4ff1-ac85-65f0a20cbdcf\") " pod="openstack/cinder-api-0" Oct 01 09:53:29 crc kubenswrapper[4787]: I1001 09:53:29.002416 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a05bd18f-4272-4ff1-ac85-65f0a20cbdcf-scripts\") pod \"cinder-api-0\" (UID: \"a05bd18f-4272-4ff1-ac85-65f0a20cbdcf\") " pod="openstack/cinder-api-0" Oct 01 09:53:29 crc kubenswrapper[4787]: I1001 09:53:29.002544 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a05bd18f-4272-4ff1-ac85-65f0a20cbdcf-config-data-custom\") pod \"cinder-api-0\" (UID: \"a05bd18f-4272-4ff1-ac85-65f0a20cbdcf\") " pod="openstack/cinder-api-0" Oct 01 09:53:29 crc kubenswrapper[4787]: I1001 09:53:29.104316 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a05bd18f-4272-4ff1-ac85-65f0a20cbdcf-config-data-custom\") pod \"cinder-api-0\" (UID: \"a05bd18f-4272-4ff1-ac85-65f0a20cbdcf\") " pod="openstack/cinder-api-0" Oct 01 09:53:29 crc kubenswrapper[4787]: I1001 09:53:29.104396 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a05bd18f-4272-4ff1-ac85-65f0a20cbdcf-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"a05bd18f-4272-4ff1-ac85-65f0a20cbdcf\") " pod="openstack/cinder-api-0" Oct 01 09:53:29 crc kubenswrapper[4787]: I1001 09:53:29.104426 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a05bd18f-4272-4ff1-ac85-65f0a20cbdcf-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a05bd18f-4272-4ff1-ac85-65f0a20cbdcf\") " pod="openstack/cinder-api-0" Oct 01 09:53:29 crc kubenswrapper[4787]: I1001 09:53:29.104450 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a05bd18f-4272-4ff1-ac85-65f0a20cbdcf-public-tls-certs\") pod \"cinder-api-0\" (UID: \"a05bd18f-4272-4ff1-ac85-65f0a20cbdcf\") " pod="openstack/cinder-api-0" Oct 01 09:53:29 crc kubenswrapper[4787]: I1001 09:53:29.104474 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a05bd18f-4272-4ff1-ac85-65f0a20cbdcf-config-data\") pod \"cinder-api-0\" (UID: \"a05bd18f-4272-4ff1-ac85-65f0a20cbdcf\") " pod="openstack/cinder-api-0" Oct 01 09:53:29 crc kubenswrapper[4787]: I1001 09:53:29.104501 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a05bd18f-4272-4ff1-ac85-65f0a20cbdcf-logs\") pod \"cinder-api-0\" (UID: \"a05bd18f-4272-4ff1-ac85-65f0a20cbdcf\") " pod="openstack/cinder-api-0" Oct 01 09:53:29 crc kubenswrapper[4787]: I1001 09:53:29.104526 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a05bd18f-4272-4ff1-ac85-65f0a20cbdcf-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a05bd18f-4272-4ff1-ac85-65f0a20cbdcf\") " pod="openstack/cinder-api-0" Oct 01 09:53:29 crc kubenswrapper[4787]: I1001 09:53:29.104545 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jg5kt\" (UniqueName: \"kubernetes.io/projected/a05bd18f-4272-4ff1-ac85-65f0a20cbdcf-kube-api-access-jg5kt\") pod \"cinder-api-0\" (UID: \"a05bd18f-4272-4ff1-ac85-65f0a20cbdcf\") " pod="openstack/cinder-api-0" Oct 01 09:53:29 crc kubenswrapper[4787]: I1001 09:53:29.104584 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a05bd18f-4272-4ff1-ac85-65f0a20cbdcf-scripts\") pod \"cinder-api-0\" (UID: \"a05bd18f-4272-4ff1-ac85-65f0a20cbdcf\") " pod="openstack/cinder-api-0" Oct 01 09:53:29 crc kubenswrapper[4787]: I1001 09:53:29.108651 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a05bd18f-4272-4ff1-ac85-65f0a20cbdcf-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a05bd18f-4272-4ff1-ac85-65f0a20cbdcf\") " pod="openstack/cinder-api-0" Oct 01 09:53:29 crc kubenswrapper[4787]: I1001 09:53:29.109701 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a05bd18f-4272-4ff1-ac85-65f0a20cbdcf-scripts\") pod \"cinder-api-0\" (UID: \"a05bd18f-4272-4ff1-ac85-65f0a20cbdcf\") " pod="openstack/cinder-api-0" Oct 01 09:53:29 crc kubenswrapper[4787]: I1001 09:53:29.110141 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a05bd18f-4272-4ff1-ac85-65f0a20cbdcf-logs\") pod \"cinder-api-0\" (UID: \"a05bd18f-4272-4ff1-ac85-65f0a20cbdcf\") " pod="openstack/cinder-api-0" Oct 01 09:53:29 crc kubenswrapper[4787]: I1001 09:53:29.112613 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a05bd18f-4272-4ff1-ac85-65f0a20cbdcf-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a05bd18f-4272-4ff1-ac85-65f0a20cbdcf\") " pod="openstack/cinder-api-0" Oct 01 09:53:29 crc kubenswrapper[4787]: I1001 09:53:29.116010 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a05bd18f-4272-4ff1-ac85-65f0a20cbdcf-config-data-custom\") pod \"cinder-api-0\" (UID: \"a05bd18f-4272-4ff1-ac85-65f0a20cbdcf\") " pod="openstack/cinder-api-0" Oct 01 09:53:29 crc kubenswrapper[4787]: I1001 09:53:29.122974 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a05bd18f-4272-4ff1-ac85-65f0a20cbdcf-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"a05bd18f-4272-4ff1-ac85-65f0a20cbdcf\") " pod="openstack/cinder-api-0" Oct 01 09:53:29 crc kubenswrapper[4787]: I1001 09:53:29.131764 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jg5kt\" (UniqueName: \"kubernetes.io/projected/a05bd18f-4272-4ff1-ac85-65f0a20cbdcf-kube-api-access-jg5kt\") pod \"cinder-api-0\" (UID: \"a05bd18f-4272-4ff1-ac85-65f0a20cbdcf\") " pod="openstack/cinder-api-0" Oct 01 09:53:29 crc kubenswrapper[4787]: I1001 09:53:29.132144 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a05bd18f-4272-4ff1-ac85-65f0a20cbdcf-config-data\") pod \"cinder-api-0\" (UID: \"a05bd18f-4272-4ff1-ac85-65f0a20cbdcf\") " pod="openstack/cinder-api-0" Oct 01 09:53:29 crc kubenswrapper[4787]: I1001 09:53:29.138998 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a05bd18f-4272-4ff1-ac85-65f0a20cbdcf-public-tls-certs\") pod \"cinder-api-0\" (UID: \"a05bd18f-4272-4ff1-ac85-65f0a20cbdcf\") " pod="openstack/cinder-api-0" Oct 01 09:53:29 crc kubenswrapper[4787]: I1001 09:53:29.298232 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 01 09:53:29 crc kubenswrapper[4787]: I1001 09:53:29.632762 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d117eceb-fcde-43be-b1be-d6acbfe6c0ac","Type":"ContainerStarted","Data":"eabce77fcf8ed644a1c6ec605e4e3a3016553d06edbb8eac83e0be38b366efcf"} Oct 01 09:53:29 crc kubenswrapper[4787]: I1001 09:53:29.633707 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 01 09:53:29 crc kubenswrapper[4787]: I1001 09:53:29.644396 4787 generic.go:334] "Generic (PLEG): container finished" podID="0d768476-fd86-440c-9c5d-dd62b6cbcaa5" containerID="39c83d707b3a7ad213890cc04ebe3e5334895c38861f85f05c13ed7f5a73cb19" exitCode=0 Oct 01 09:53:29 crc kubenswrapper[4787]: I1001 09:53:29.646717 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-400d-account-create-mtvjl" event={"ID":"0d768476-fd86-440c-9c5d-dd62b6cbcaa5","Type":"ContainerDied","Data":"39c83d707b3a7ad213890cc04ebe3e5334895c38861f85f05c13ed7f5a73cb19"} Oct 01 09:53:29 crc kubenswrapper[4787]: I1001 09:53:29.646789 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-400d-account-create-mtvjl" event={"ID":"0d768476-fd86-440c-9c5d-dd62b6cbcaa5","Type":"ContainerStarted","Data":"56c3daa8184cadd6c5c67fc2582acd83288070a3c247066f29532439e5390de5"} Oct 01 09:53:29 crc kubenswrapper[4787]: I1001 09:53:29.672306 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.932832644 podStartE2EDuration="7.67226818s" podCreationTimestamp="2025-10-01 09:53:22 +0000 UTC" firstStartedPulling="2025-10-01 09:53:24.330128373 +0000 UTC m=+1036.445272530" lastFinishedPulling="2025-10-01 09:53:29.069563909 +0000 UTC m=+1041.184708066" observedRunningTime="2025-10-01 09:53:29.660652825 +0000 UTC m=+1041.775796992" watchObservedRunningTime="2025-10-01 09:53:29.67226818 +0000 UTC m=+1041.787412337" Oct 01 09:53:29 crc kubenswrapper[4787]: I1001 09:53:29.809735 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 01 09:53:30 crc kubenswrapper[4787]: I1001 09:53:30.343744 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-58d9cd74f7-86sts" Oct 01 09:53:30 crc kubenswrapper[4787]: I1001 09:53:30.479467 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-858fdb94fd-6tpq6"] Oct 01 09:53:30 crc kubenswrapper[4787]: I1001 09:53:30.480467 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-858fdb94fd-6tpq6" podUID="0ac97b66-7233-4fcc-9b0f-856a67a62b4d" containerName="neutron-api" containerID="cri-o://09101ceda6bf2ba07365f993acdcbbb18dd498f848daa46e32c04bb05065a77b" gracePeriod=30 Oct 01 09:53:30 crc kubenswrapper[4787]: I1001 09:53:30.480716 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-858fdb94fd-6tpq6" podUID="0ac97b66-7233-4fcc-9b0f-856a67a62b4d" containerName="neutron-httpd" containerID="cri-o://2b5867e80774c39c9e0b986b908a0d57bfbab946cc4e15e39311fc5003ff06a0" gracePeriod=30 Oct 01 09:53:30 crc kubenswrapper[4787]: I1001 09:53:30.541294 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="165b7740-cef4-45af-87d1-901b882b8dbb" path="/var/lib/kubelet/pods/165b7740-cef4-45af-87d1-901b882b8dbb/volumes" Oct 01 09:53:30 crc kubenswrapper[4787]: I1001 09:53:30.542225 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7983e213-c317-42b2-8944-85bdb78dac8f" path="/var/lib/kubelet/pods/7983e213-c317-42b2-8944-85bdb78dac8f/volumes" Oct 01 09:53:30 crc kubenswrapper[4787]: I1001 09:53:30.676371 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a05bd18f-4272-4ff1-ac85-65f0a20cbdcf","Type":"ContainerStarted","Data":"22c86117cce6bc8acd93ebb6ec605819bc0131a207b8f6c5bd66971abc71f7ac"} Oct 01 09:53:30 crc kubenswrapper[4787]: I1001 09:53:30.676614 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a05bd18f-4272-4ff1-ac85-65f0a20cbdcf","Type":"ContainerStarted","Data":"323eba21628ae232ebb9f200232f13b4ceaf63e4858190624d0718a657a98c2d"} Oct 01 09:53:31 crc kubenswrapper[4787]: I1001 09:53:31.091118 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-400d-account-create-mtvjl" Oct 01 09:53:31 crc kubenswrapper[4787]: I1001 09:53:31.261513 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wl7qz\" (UniqueName: \"kubernetes.io/projected/0d768476-fd86-440c-9c5d-dd62b6cbcaa5-kube-api-access-wl7qz\") pod \"0d768476-fd86-440c-9c5d-dd62b6cbcaa5\" (UID: \"0d768476-fd86-440c-9c5d-dd62b6cbcaa5\") " Oct 01 09:53:31 crc kubenswrapper[4787]: I1001 09:53:31.283483 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d768476-fd86-440c-9c5d-dd62b6cbcaa5-kube-api-access-wl7qz" (OuterVolumeSpecName: "kube-api-access-wl7qz") pod "0d768476-fd86-440c-9c5d-dd62b6cbcaa5" (UID: "0d768476-fd86-440c-9c5d-dd62b6cbcaa5"). InnerVolumeSpecName "kube-api-access-wl7qz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:53:31 crc kubenswrapper[4787]: I1001 09:53:31.364048 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wl7qz\" (UniqueName: \"kubernetes.io/projected/0d768476-fd86-440c-9c5d-dd62b6cbcaa5-kube-api-access-wl7qz\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:31 crc kubenswrapper[4787]: I1001 09:53:31.689574 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a05bd18f-4272-4ff1-ac85-65f0a20cbdcf","Type":"ContainerStarted","Data":"5e66d81dee83d75d703dede619e840436e87b69902cc2038764b3be7179f6fa0"} Oct 01 09:53:31 crc kubenswrapper[4787]: I1001 09:53:31.689715 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 01 09:53:31 crc kubenswrapper[4787]: I1001 09:53:31.692600 4787 generic.go:334] "Generic (PLEG): container finished" podID="0ac97b66-7233-4fcc-9b0f-856a67a62b4d" containerID="2b5867e80774c39c9e0b986b908a0d57bfbab946cc4e15e39311fc5003ff06a0" exitCode=0 Oct 01 09:53:31 crc kubenswrapper[4787]: I1001 09:53:31.692649 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-858fdb94fd-6tpq6" event={"ID":"0ac97b66-7233-4fcc-9b0f-856a67a62b4d","Type":"ContainerDied","Data":"2b5867e80774c39c9e0b986b908a0d57bfbab946cc4e15e39311fc5003ff06a0"} Oct 01 09:53:31 crc kubenswrapper[4787]: I1001 09:53:31.695667 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-400d-account-create-mtvjl" event={"ID":"0d768476-fd86-440c-9c5d-dd62b6cbcaa5","Type":"ContainerDied","Data":"56c3daa8184cadd6c5c67fc2582acd83288070a3c247066f29532439e5390de5"} Oct 01 09:53:31 crc kubenswrapper[4787]: I1001 09:53:31.695704 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="56c3daa8184cadd6c5c67fc2582acd83288070a3c247066f29532439e5390de5" Oct 01 09:53:31 crc kubenswrapper[4787]: I1001 09:53:31.695740 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-400d-account-create-mtvjl" Oct 01 09:53:31 crc kubenswrapper[4787]: I1001 09:53:31.708627 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.708605941 podStartE2EDuration="3.708605941s" podCreationTimestamp="2025-10-01 09:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:53:31.707295068 +0000 UTC m=+1043.822439235" watchObservedRunningTime="2025-10-01 09:53:31.708605941 +0000 UTC m=+1043.823750088" Oct 01 09:53:32 crc kubenswrapper[4787]: I1001 09:53:32.153501 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:53:32 crc kubenswrapper[4787]: I1001 09:53:32.153801 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d117eceb-fcde-43be-b1be-d6acbfe6c0ac" containerName="ceilometer-central-agent" containerID="cri-o://082bb82e79dc55abe18103279f1bf58f67b45b73bb7d96f7933aa0d852a259f5" gracePeriod=30 Oct 01 09:53:32 crc kubenswrapper[4787]: I1001 09:53:32.153888 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d117eceb-fcde-43be-b1be-d6acbfe6c0ac" containerName="sg-core" containerID="cri-o://831afab4a31e797b058e98f7db738f32e779ff985867fb2ce615b1d9f6bbe02a" gracePeriod=30 Oct 01 09:53:32 crc kubenswrapper[4787]: I1001 09:53:32.153982 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d117eceb-fcde-43be-b1be-d6acbfe6c0ac" containerName="ceilometer-notification-agent" containerID="cri-o://f8660db311600b14f532f6d7a168f19b043ddc07ac600d493be40f786ac58844" gracePeriod=30 Oct 01 09:53:32 crc kubenswrapper[4787]: I1001 09:53:32.154090 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d117eceb-fcde-43be-b1be-d6acbfe6c0ac" containerName="proxy-httpd" containerID="cri-o://eabce77fcf8ed644a1c6ec605e4e3a3016553d06edbb8eac83e0be38b366efcf" gracePeriod=30 Oct 01 09:53:32 crc kubenswrapper[4787]: I1001 09:53:32.758655 4787 generic.go:334] "Generic (PLEG): container finished" podID="d117eceb-fcde-43be-b1be-d6acbfe6c0ac" containerID="eabce77fcf8ed644a1c6ec605e4e3a3016553d06edbb8eac83e0be38b366efcf" exitCode=0 Oct 01 09:53:32 crc kubenswrapper[4787]: I1001 09:53:32.759142 4787 generic.go:334] "Generic (PLEG): container finished" podID="d117eceb-fcde-43be-b1be-d6acbfe6c0ac" containerID="831afab4a31e797b058e98f7db738f32e779ff985867fb2ce615b1d9f6bbe02a" exitCode=2 Oct 01 09:53:32 crc kubenswrapper[4787]: I1001 09:53:32.759154 4787 generic.go:334] "Generic (PLEG): container finished" podID="d117eceb-fcde-43be-b1be-d6acbfe6c0ac" containerID="f8660db311600b14f532f6d7a168f19b043ddc07ac600d493be40f786ac58844" exitCode=0 Oct 01 09:53:32 crc kubenswrapper[4787]: I1001 09:53:32.759164 4787 generic.go:334] "Generic (PLEG): container finished" podID="d117eceb-fcde-43be-b1be-d6acbfe6c0ac" containerID="082bb82e79dc55abe18103279f1bf58f67b45b73bb7d96f7933aa0d852a259f5" exitCode=0 Oct 01 09:53:32 crc kubenswrapper[4787]: I1001 09:53:32.759460 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d117eceb-fcde-43be-b1be-d6acbfe6c0ac","Type":"ContainerDied","Data":"eabce77fcf8ed644a1c6ec605e4e3a3016553d06edbb8eac83e0be38b366efcf"} Oct 01 09:53:32 crc kubenswrapper[4787]: I1001 09:53:32.759504 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d117eceb-fcde-43be-b1be-d6acbfe6c0ac","Type":"ContainerDied","Data":"831afab4a31e797b058e98f7db738f32e779ff985867fb2ce615b1d9f6bbe02a"} Oct 01 09:53:32 crc kubenswrapper[4787]: I1001 09:53:32.759519 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d117eceb-fcde-43be-b1be-d6acbfe6c0ac","Type":"ContainerDied","Data":"f8660db311600b14f532f6d7a168f19b043ddc07ac600d493be40f786ac58844"} Oct 01 09:53:32 crc kubenswrapper[4787]: I1001 09:53:32.759547 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d117eceb-fcde-43be-b1be-d6acbfe6c0ac","Type":"ContainerDied","Data":"082bb82e79dc55abe18103279f1bf58f67b45b73bb7d96f7933aa0d852a259f5"} Oct 01 09:53:32 crc kubenswrapper[4787]: I1001 09:53:32.992760 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.126852 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.199811 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-config-data\") pod \"d117eceb-fcde-43be-b1be-d6acbfe6c0ac\" (UID: \"d117eceb-fcde-43be-b1be-d6acbfe6c0ac\") " Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.199882 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lw8dm\" (UniqueName: \"kubernetes.io/projected/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-kube-api-access-lw8dm\") pod \"d117eceb-fcde-43be-b1be-d6acbfe6c0ac\" (UID: \"d117eceb-fcde-43be-b1be-d6acbfe6c0ac\") " Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.199922 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-scripts\") pod \"d117eceb-fcde-43be-b1be-d6acbfe6c0ac\" (UID: \"d117eceb-fcde-43be-b1be-d6acbfe6c0ac\") " Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.199946 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-log-httpd\") pod \"d117eceb-fcde-43be-b1be-d6acbfe6c0ac\" (UID: \"d117eceb-fcde-43be-b1be-d6acbfe6c0ac\") " Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.199986 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-sg-core-conf-yaml\") pod \"d117eceb-fcde-43be-b1be-d6acbfe6c0ac\" (UID: \"d117eceb-fcde-43be-b1be-d6acbfe6c0ac\") " Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.200048 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-combined-ca-bundle\") pod \"d117eceb-fcde-43be-b1be-d6acbfe6c0ac\" (UID: \"d117eceb-fcde-43be-b1be-d6acbfe6c0ac\") " Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.200069 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-run-httpd\") pod \"d117eceb-fcde-43be-b1be-d6acbfe6c0ac\" (UID: \"d117eceb-fcde-43be-b1be-d6acbfe6c0ac\") " Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.201253 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d117eceb-fcde-43be-b1be-d6acbfe6c0ac" (UID: "d117eceb-fcde-43be-b1be-d6acbfe6c0ac"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.201578 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d117eceb-fcde-43be-b1be-d6acbfe6c0ac" (UID: "d117eceb-fcde-43be-b1be-d6acbfe6c0ac"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.212502 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-scripts" (OuterVolumeSpecName: "scripts") pod "d117eceb-fcde-43be-b1be-d6acbfe6c0ac" (UID: "d117eceb-fcde-43be-b1be-d6acbfe6c0ac"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.219347 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-kube-api-access-lw8dm" (OuterVolumeSpecName: "kube-api-access-lw8dm") pod "d117eceb-fcde-43be-b1be-d6acbfe6c0ac" (UID: "d117eceb-fcde-43be-b1be-d6acbfe6c0ac"). InnerVolumeSpecName "kube-api-access-lw8dm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.264206 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d117eceb-fcde-43be-b1be-d6acbfe6c0ac" (UID: "d117eceb-fcde-43be-b1be-d6acbfe6c0ac"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.269409 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7bd7989875-t7k69" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.302509 4787 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.302537 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lw8dm\" (UniqueName: \"kubernetes.io/projected/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-kube-api-access-lw8dm\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.302546 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.302556 4787 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.302564 4787 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.335596 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d117eceb-fcde-43be-b1be-d6acbfe6c0ac" (UID: "d117eceb-fcde-43be-b1be-d6acbfe6c0ac"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.363372 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f4d8965b5-7dw2j"] Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.363628 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-f4d8965b5-7dw2j" podUID="17fa3c34-c304-4972-b023-879f3ee28416" containerName="dnsmasq-dns" containerID="cri-o://1f31d9c2da9765e7b68759a10bde045a5fa9ebda788415b3850019ea6831e3eb" gracePeriod=10 Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.404181 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.466542 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-config-data" (OuterVolumeSpecName: "config-data") pod "d117eceb-fcde-43be-b1be-d6acbfe6c0ac" (UID: "d117eceb-fcde-43be-b1be-d6acbfe6c0ac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.506517 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d117eceb-fcde-43be-b1be-d6acbfe6c0ac-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.595195 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.595651 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.649560 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.654314 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.760780 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.778102 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d117eceb-fcde-43be-b1be-d6acbfe6c0ac","Type":"ContainerDied","Data":"4a202fc268274228b45839320fe265d7a6349eebc719049529e5e1a1cabf4e35"} Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.778178 4787 scope.go:117] "RemoveContainer" containerID="eabce77fcf8ed644a1c6ec605e4e3a3016553d06edbb8eac83e0be38b366efcf" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.778396 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.787983 4787 generic.go:334] "Generic (PLEG): container finished" podID="17fa3c34-c304-4972-b023-879f3ee28416" containerID="1f31d9c2da9765e7b68759a10bde045a5fa9ebda788415b3850019ea6831e3eb" exitCode=0 Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.788061 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f4d8965b5-7dw2j" event={"ID":"17fa3c34-c304-4972-b023-879f3ee28416","Type":"ContainerDied","Data":"1f31d9c2da9765e7b68759a10bde045a5fa9ebda788415b3850019ea6831e3eb"} Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.789627 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.789644 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.868737 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.882602 4787 scope.go:117] "RemoveContainer" containerID="831afab4a31e797b058e98f7db738f32e779ff985867fb2ce615b1d9f6bbe02a" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.885216 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.903195 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.907774 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f4d8965b5-7dw2j" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.919710 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:53:33 crc kubenswrapper[4787]: E1001 09:53:33.920141 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d117eceb-fcde-43be-b1be-d6acbfe6c0ac" containerName="ceilometer-central-agent" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.920156 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="d117eceb-fcde-43be-b1be-d6acbfe6c0ac" containerName="ceilometer-central-agent" Oct 01 09:53:33 crc kubenswrapper[4787]: E1001 09:53:33.920171 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d117eceb-fcde-43be-b1be-d6acbfe6c0ac" containerName="ceilometer-notification-agent" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.920177 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="d117eceb-fcde-43be-b1be-d6acbfe6c0ac" containerName="ceilometer-notification-agent" Oct 01 09:53:33 crc kubenswrapper[4787]: E1001 09:53:33.920199 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d768476-fd86-440c-9c5d-dd62b6cbcaa5" containerName="mariadb-account-create" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.920205 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d768476-fd86-440c-9c5d-dd62b6cbcaa5" containerName="mariadb-account-create" Oct 01 09:53:33 crc kubenswrapper[4787]: E1001 09:53:33.920221 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17fa3c34-c304-4972-b023-879f3ee28416" containerName="dnsmasq-dns" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.920227 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="17fa3c34-c304-4972-b023-879f3ee28416" containerName="dnsmasq-dns" Oct 01 09:53:33 crc kubenswrapper[4787]: E1001 09:53:33.920236 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d117eceb-fcde-43be-b1be-d6acbfe6c0ac" containerName="proxy-httpd" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.920243 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="d117eceb-fcde-43be-b1be-d6acbfe6c0ac" containerName="proxy-httpd" Oct 01 09:53:33 crc kubenswrapper[4787]: E1001 09:53:33.920261 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17fa3c34-c304-4972-b023-879f3ee28416" containerName="init" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.920268 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="17fa3c34-c304-4972-b023-879f3ee28416" containerName="init" Oct 01 09:53:33 crc kubenswrapper[4787]: E1001 09:53:33.920281 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d117eceb-fcde-43be-b1be-d6acbfe6c0ac" containerName="sg-core" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.920289 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="d117eceb-fcde-43be-b1be-d6acbfe6c0ac" containerName="sg-core" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.920467 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d768476-fd86-440c-9c5d-dd62b6cbcaa5" containerName="mariadb-account-create" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.920477 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="d117eceb-fcde-43be-b1be-d6acbfe6c0ac" containerName="proxy-httpd" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.920489 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="d117eceb-fcde-43be-b1be-d6acbfe6c0ac" containerName="ceilometer-central-agent" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.920499 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="17fa3c34-c304-4972-b023-879f3ee28416" containerName="dnsmasq-dns" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.920509 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="d117eceb-fcde-43be-b1be-d6acbfe6c0ac" containerName="sg-core" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.920528 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="d117eceb-fcde-43be-b1be-d6acbfe6c0ac" containerName="ceilometer-notification-agent" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.932053 4787 scope.go:117] "RemoveContainer" containerID="f8660db311600b14f532f6d7a168f19b043ddc07ac600d493be40f786ac58844" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.944942 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.944986 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.945003 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.945292 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.960102 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.960845 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.991569 4787 scope.go:117] "RemoveContainer" containerID="082bb82e79dc55abe18103279f1bf58f67b45b73bb7d96f7933aa0d852a259f5" Oct 01 09:53:33 crc kubenswrapper[4787]: I1001 09:53:33.996993 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.019323 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17fa3c34-c304-4972-b023-879f3ee28416-ovsdbserver-nb\") pod \"17fa3c34-c304-4972-b023-879f3ee28416\" (UID: \"17fa3c34-c304-4972-b023-879f3ee28416\") " Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.019422 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17fa3c34-c304-4972-b023-879f3ee28416-dns-svc\") pod \"17fa3c34-c304-4972-b023-879f3ee28416\" (UID: \"17fa3c34-c304-4972-b023-879f3ee28416\") " Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.019471 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17fa3c34-c304-4972-b023-879f3ee28416-ovsdbserver-sb\") pod \"17fa3c34-c304-4972-b023-879f3ee28416\" (UID: \"17fa3c34-c304-4972-b023-879f3ee28416\") " Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.019551 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17fa3c34-c304-4972-b023-879f3ee28416-config\") pod \"17fa3c34-c304-4972-b023-879f3ee28416\" (UID: \"17fa3c34-c304-4972-b023-879f3ee28416\") " Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.020727 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjgm2\" (UniqueName: \"kubernetes.io/projected/17fa3c34-c304-4972-b023-879f3ee28416-kube-api-access-sjgm2\") pod \"17fa3c34-c304-4972-b023-879f3ee28416\" (UID: \"17fa3c34-c304-4972-b023-879f3ee28416\") " Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.020799 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/17fa3c34-c304-4972-b023-879f3ee28416-dns-swift-storage-0\") pod \"17fa3c34-c304-4972-b023-879f3ee28416\" (UID: \"17fa3c34-c304-4972-b023-879f3ee28416\") " Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.021044 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14c412a3-bcdc-4a19-b769-310820125df0-log-httpd\") pod \"ceilometer-0\" (UID: \"14c412a3-bcdc-4a19-b769-310820125df0\") " pod="openstack/ceilometer-0" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.021120 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/14c412a3-bcdc-4a19-b769-310820125df0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"14c412a3-bcdc-4a19-b769-310820125df0\") " pod="openstack/ceilometer-0" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.021160 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14c412a3-bcdc-4a19-b769-310820125df0-run-httpd\") pod \"ceilometer-0\" (UID: \"14c412a3-bcdc-4a19-b769-310820125df0\") " pod="openstack/ceilometer-0" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.021204 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14c412a3-bcdc-4a19-b769-310820125df0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"14c412a3-bcdc-4a19-b769-310820125df0\") " pod="openstack/ceilometer-0" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.021223 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14c412a3-bcdc-4a19-b769-310820125df0-config-data\") pod \"ceilometer-0\" (UID: \"14c412a3-bcdc-4a19-b769-310820125df0\") " pod="openstack/ceilometer-0" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.021240 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbxtt\" (UniqueName: \"kubernetes.io/projected/14c412a3-bcdc-4a19-b769-310820125df0-kube-api-access-zbxtt\") pod \"ceilometer-0\" (UID: \"14c412a3-bcdc-4a19-b769-310820125df0\") " pod="openstack/ceilometer-0" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.021372 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14c412a3-bcdc-4a19-b769-310820125df0-scripts\") pod \"ceilometer-0\" (UID: \"14c412a3-bcdc-4a19-b769-310820125df0\") " pod="openstack/ceilometer-0" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.034000 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.038889 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17fa3c34-c304-4972-b023-879f3ee28416-kube-api-access-sjgm2" (OuterVolumeSpecName: "kube-api-access-sjgm2") pod "17fa3c34-c304-4972-b023-879f3ee28416" (UID: "17fa3c34-c304-4972-b023-879f3ee28416"). InnerVolumeSpecName "kube-api-access-sjgm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.122292 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17fa3c34-c304-4972-b023-879f3ee28416-config" (OuterVolumeSpecName: "config") pod "17fa3c34-c304-4972-b023-879f3ee28416" (UID: "17fa3c34-c304-4972-b023-879f3ee28416"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.123324 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/14c412a3-bcdc-4a19-b769-310820125df0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"14c412a3-bcdc-4a19-b769-310820125df0\") " pod="openstack/ceilometer-0" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.123366 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14c412a3-bcdc-4a19-b769-310820125df0-run-httpd\") pod \"ceilometer-0\" (UID: \"14c412a3-bcdc-4a19-b769-310820125df0\") " pod="openstack/ceilometer-0" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.123406 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14c412a3-bcdc-4a19-b769-310820125df0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"14c412a3-bcdc-4a19-b769-310820125df0\") " pod="openstack/ceilometer-0" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.123423 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14c412a3-bcdc-4a19-b769-310820125df0-config-data\") pod \"ceilometer-0\" (UID: \"14c412a3-bcdc-4a19-b769-310820125df0\") " pod="openstack/ceilometer-0" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.123439 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbxtt\" (UniqueName: \"kubernetes.io/projected/14c412a3-bcdc-4a19-b769-310820125df0-kube-api-access-zbxtt\") pod \"ceilometer-0\" (UID: \"14c412a3-bcdc-4a19-b769-310820125df0\") " pod="openstack/ceilometer-0" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.123510 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14c412a3-bcdc-4a19-b769-310820125df0-scripts\") pod \"ceilometer-0\" (UID: \"14c412a3-bcdc-4a19-b769-310820125df0\") " pod="openstack/ceilometer-0" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.123555 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14c412a3-bcdc-4a19-b769-310820125df0-log-httpd\") pod \"ceilometer-0\" (UID: \"14c412a3-bcdc-4a19-b769-310820125df0\") " pod="openstack/ceilometer-0" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.123620 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17fa3c34-c304-4972-b023-879f3ee28416-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.123634 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjgm2\" (UniqueName: \"kubernetes.io/projected/17fa3c34-c304-4972-b023-879f3ee28416-kube-api-access-sjgm2\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.123923 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14c412a3-bcdc-4a19-b769-310820125df0-log-httpd\") pod \"ceilometer-0\" (UID: \"14c412a3-bcdc-4a19-b769-310820125df0\") " pod="openstack/ceilometer-0" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.125386 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14c412a3-bcdc-4a19-b769-310820125df0-run-httpd\") pod \"ceilometer-0\" (UID: \"14c412a3-bcdc-4a19-b769-310820125df0\") " pod="openstack/ceilometer-0" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.131198 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/14c412a3-bcdc-4a19-b769-310820125df0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"14c412a3-bcdc-4a19-b769-310820125df0\") " pod="openstack/ceilometer-0" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.135207 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14c412a3-bcdc-4a19-b769-310820125df0-scripts\") pod \"ceilometer-0\" (UID: \"14c412a3-bcdc-4a19-b769-310820125df0\") " pod="openstack/ceilometer-0" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.138709 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14c412a3-bcdc-4a19-b769-310820125df0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"14c412a3-bcdc-4a19-b769-310820125df0\") " pod="openstack/ceilometer-0" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.138851 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14c412a3-bcdc-4a19-b769-310820125df0-config-data\") pod \"ceilometer-0\" (UID: \"14c412a3-bcdc-4a19-b769-310820125df0\") " pod="openstack/ceilometer-0" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.143028 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17fa3c34-c304-4972-b023-879f3ee28416-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "17fa3c34-c304-4972-b023-879f3ee28416" (UID: "17fa3c34-c304-4972-b023-879f3ee28416"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.143058 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbxtt\" (UniqueName: \"kubernetes.io/projected/14c412a3-bcdc-4a19-b769-310820125df0-kube-api-access-zbxtt\") pod \"ceilometer-0\" (UID: \"14c412a3-bcdc-4a19-b769-310820125df0\") " pod="openstack/ceilometer-0" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.162797 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17fa3c34-c304-4972-b023-879f3ee28416-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "17fa3c34-c304-4972-b023-879f3ee28416" (UID: "17fa3c34-c304-4972-b023-879f3ee28416"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.172755 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17fa3c34-c304-4972-b023-879f3ee28416-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "17fa3c34-c304-4972-b023-879f3ee28416" (UID: "17fa3c34-c304-4972-b023-879f3ee28416"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.180468 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17fa3c34-c304-4972-b023-879f3ee28416-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "17fa3c34-c304-4972-b023-879f3ee28416" (UID: "17fa3c34-c304-4972-b023-879f3ee28416"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.226118 4787 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17fa3c34-c304-4972-b023-879f3ee28416-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.226158 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17fa3c34-c304-4972-b023-879f3ee28416-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.226172 4787 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/17fa3c34-c304-4972-b023-879f3ee28416-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.226192 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17fa3c34-c304-4972-b023-879f3ee28416-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.281415 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.538056 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d117eceb-fcde-43be-b1be-d6acbfe6c0ac" path="/var/lib/kubelet/pods/d117eceb-fcde-43be-b1be-d6acbfe6c0ac/volumes" Oct 01 09:53:34 crc kubenswrapper[4787]: E1001 09:53:34.795340 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17fa3c34_c304_4972_b023_879f3ee28416.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17fa3c34_c304_4972_b023_879f3ee28416.slice/crio-a12d02ab702e8853288c4fdf23cb0e7cbf037021907fee6fea74d55f3f557d66\": RecentStats: unable to find data in memory cache]" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.807046 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14c412a3-bcdc-4a19-b769-310820125df0","Type":"ContainerStarted","Data":"c5bfecd60b3009268f3c0a020b91d26523d48fc5748cc805fb4bad922ed5e8f5"} Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.812202 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.815629 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f4d8965b5-7dw2j" event={"ID":"17fa3c34-c304-4972-b023-879f3ee28416","Type":"ContainerDied","Data":"a12d02ab702e8853288c4fdf23cb0e7cbf037021907fee6fea74d55f3f557d66"} Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.815681 4787 scope.go:117] "RemoveContainer" containerID="1f31d9c2da9765e7b68759a10bde045a5fa9ebda788415b3850019ea6831e3eb" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.815878 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f4d8965b5-7dw2j" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.821740 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="6eff6348-9415-4d61-a405-0a6ac27869d4" containerName="cinder-scheduler" containerID="cri-o://5784ec8cb42dd5d8922a1fcfb5e4e0b56515623a7de4cb21e67610f768fd5bc3" gracePeriod=30 Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.821987 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="6eff6348-9415-4d61-a405-0a6ac27869d4" containerName="probe" containerID="cri-o://c9ecc450edb50582561268ad6d9c1a2bd4afd58b50ac192de75e9b62a534e973" gracePeriod=30 Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.822443 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.822473 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.854529 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f4d8965b5-7dw2j"] Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.860706 4787 scope.go:117] "RemoveContainer" containerID="f931d2dda239ceee3044a948206acf3182370a53555e71768aa5a54e24c15e01" Oct 01 09:53:34 crc kubenswrapper[4787]: I1001 09:53:34.863190 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f4d8965b5-7dw2j"] Oct 01 09:53:35 crc kubenswrapper[4787]: I1001 09:53:35.546850 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:53:35 crc kubenswrapper[4787]: I1001 09:53:35.846724 4787 generic.go:334] "Generic (PLEG): container finished" podID="6eff6348-9415-4d61-a405-0a6ac27869d4" containerID="5784ec8cb42dd5d8922a1fcfb5e4e0b56515623a7de4cb21e67610f768fd5bc3" exitCode=0 Oct 01 09:53:35 crc kubenswrapper[4787]: I1001 09:53:35.847099 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6eff6348-9415-4d61-a405-0a6ac27869d4","Type":"ContainerDied","Data":"5784ec8cb42dd5d8922a1fcfb5e4e0b56515623a7de4cb21e67610f768fd5bc3"} Oct 01 09:53:35 crc kubenswrapper[4787]: I1001 09:53:35.858309 4787 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 01 09:53:35 crc kubenswrapper[4787]: I1001 09:53:35.858525 4787 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 01 09:53:35 crc kubenswrapper[4787]: I1001 09:53:35.859101 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14c412a3-bcdc-4a19-b769-310820125df0","Type":"ContainerStarted","Data":"9fef4997b8b7315a153484f251256d79d41f4ad455d3e9db1fba2a85071cfd46"} Oct 01 09:53:36 crc kubenswrapper[4787]: I1001 09:53:36.246063 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 01 09:53:36 crc kubenswrapper[4787]: I1001 09:53:36.388223 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff6348-9415-4d61-a405-0a6ac27869d4-combined-ca-bundle\") pod \"6eff6348-9415-4d61-a405-0a6ac27869d4\" (UID: \"6eff6348-9415-4d61-a405-0a6ac27869d4\") " Oct 01 09:53:36 crc kubenswrapper[4787]: I1001 09:53:36.388341 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6eff6348-9415-4d61-a405-0a6ac27869d4-scripts\") pod \"6eff6348-9415-4d61-a405-0a6ac27869d4\" (UID: \"6eff6348-9415-4d61-a405-0a6ac27869d4\") " Oct 01 09:53:36 crc kubenswrapper[4787]: I1001 09:53:36.388498 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6eff6348-9415-4d61-a405-0a6ac27869d4-config-data\") pod \"6eff6348-9415-4d61-a405-0a6ac27869d4\" (UID: \"6eff6348-9415-4d61-a405-0a6ac27869d4\") " Oct 01 09:53:36 crc kubenswrapper[4787]: I1001 09:53:36.388528 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6eff6348-9415-4d61-a405-0a6ac27869d4-config-data-custom\") pod \"6eff6348-9415-4d61-a405-0a6ac27869d4\" (UID: \"6eff6348-9415-4d61-a405-0a6ac27869d4\") " Oct 01 09:53:36 crc kubenswrapper[4787]: I1001 09:53:36.388560 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rkzz\" (UniqueName: \"kubernetes.io/projected/6eff6348-9415-4d61-a405-0a6ac27869d4-kube-api-access-8rkzz\") pod \"6eff6348-9415-4d61-a405-0a6ac27869d4\" (UID: \"6eff6348-9415-4d61-a405-0a6ac27869d4\") " Oct 01 09:53:36 crc kubenswrapper[4787]: I1001 09:53:36.388580 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6eff6348-9415-4d61-a405-0a6ac27869d4-etc-machine-id\") pod \"6eff6348-9415-4d61-a405-0a6ac27869d4\" (UID: \"6eff6348-9415-4d61-a405-0a6ac27869d4\") " Oct 01 09:53:36 crc kubenswrapper[4787]: I1001 09:53:36.389020 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6eff6348-9415-4d61-a405-0a6ac27869d4-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "6eff6348-9415-4d61-a405-0a6ac27869d4" (UID: "6eff6348-9415-4d61-a405-0a6ac27869d4"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 09:53:36 crc kubenswrapper[4787]: I1001 09:53:36.393718 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6eff6348-9415-4d61-a405-0a6ac27869d4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "6eff6348-9415-4d61-a405-0a6ac27869d4" (UID: "6eff6348-9415-4d61-a405-0a6ac27869d4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:36 crc kubenswrapper[4787]: I1001 09:53:36.397600 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6eff6348-9415-4d61-a405-0a6ac27869d4-scripts" (OuterVolumeSpecName: "scripts") pod "6eff6348-9415-4d61-a405-0a6ac27869d4" (UID: "6eff6348-9415-4d61-a405-0a6ac27869d4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:36 crc kubenswrapper[4787]: I1001 09:53:36.406504 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6eff6348-9415-4d61-a405-0a6ac27869d4-kube-api-access-8rkzz" (OuterVolumeSpecName: "kube-api-access-8rkzz") pod "6eff6348-9415-4d61-a405-0a6ac27869d4" (UID: "6eff6348-9415-4d61-a405-0a6ac27869d4"). InnerVolumeSpecName "kube-api-access-8rkzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:53:36 crc kubenswrapper[4787]: I1001 09:53:36.491163 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6eff6348-9415-4d61-a405-0a6ac27869d4-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:36 crc kubenswrapper[4787]: I1001 09:53:36.491195 4787 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6eff6348-9415-4d61-a405-0a6ac27869d4-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:36 crc kubenswrapper[4787]: I1001 09:53:36.491205 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rkzz\" (UniqueName: \"kubernetes.io/projected/6eff6348-9415-4d61-a405-0a6ac27869d4-kube-api-access-8rkzz\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:36 crc kubenswrapper[4787]: I1001 09:53:36.491214 4787 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6eff6348-9415-4d61-a405-0a6ac27869d4-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:36 crc kubenswrapper[4787]: I1001 09:53:36.494771 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6eff6348-9415-4d61-a405-0a6ac27869d4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6eff6348-9415-4d61-a405-0a6ac27869d4" (UID: "6eff6348-9415-4d61-a405-0a6ac27869d4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:36 crc kubenswrapper[4787]: I1001 09:53:36.512490 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 01 09:53:36 crc kubenswrapper[4787]: I1001 09:53:36.512549 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 01 09:53:36 crc kubenswrapper[4787]: I1001 09:53:36.534223 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17fa3c34-c304-4972-b023-879f3ee28416" path="/var/lib/kubelet/pods/17fa3c34-c304-4972-b023-879f3ee28416/volumes" Oct 01 09:53:36 crc kubenswrapper[4787]: I1001 09:53:36.588043 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6eff6348-9415-4d61-a405-0a6ac27869d4-config-data" (OuterVolumeSpecName: "config-data") pod "6eff6348-9415-4d61-a405-0a6ac27869d4" (UID: "6eff6348-9415-4d61-a405-0a6ac27869d4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:36 crc kubenswrapper[4787]: I1001 09:53:36.593498 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff6348-9415-4d61-a405-0a6ac27869d4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:36 crc kubenswrapper[4787]: I1001 09:53:36.593532 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6eff6348-9415-4d61-a405-0a6ac27869d4-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:36 crc kubenswrapper[4787]: I1001 09:53:36.898538 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14c412a3-bcdc-4a19-b769-310820125df0","Type":"ContainerStarted","Data":"820b434e99203b0219ec476dc6a6a9fdfe43a8a8957692874fe905ef8dc37585"} Oct 01 09:53:36 crc kubenswrapper[4787]: I1001 09:53:36.905243 4787 generic.go:334] "Generic (PLEG): container finished" podID="6eff6348-9415-4d61-a405-0a6ac27869d4" containerID="c9ecc450edb50582561268ad6d9c1a2bd4afd58b50ac192de75e9b62a534e973" exitCode=0 Oct 01 09:53:36 crc kubenswrapper[4787]: I1001 09:53:36.905332 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6eff6348-9415-4d61-a405-0a6ac27869d4","Type":"ContainerDied","Data":"c9ecc450edb50582561268ad6d9c1a2bd4afd58b50ac192de75e9b62a534e973"} Oct 01 09:53:36 crc kubenswrapper[4787]: I1001 09:53:36.905368 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6eff6348-9415-4d61-a405-0a6ac27869d4","Type":"ContainerDied","Data":"1bd39e017189758cb0462d0ab61f7301e1c4322581543a1d3d63a9807fe40b93"} Oct 01 09:53:36 crc kubenswrapper[4787]: I1001 09:53:36.905390 4787 scope.go:117] "RemoveContainer" containerID="c9ecc450edb50582561268ad6d9c1a2bd4afd58b50ac192de75e9b62a534e973" Oct 01 09:53:36 crc kubenswrapper[4787]: I1001 09:53:36.905491 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 01 09:53:36 crc kubenswrapper[4787]: I1001 09:53:36.910534 4787 generic.go:334] "Generic (PLEG): container finished" podID="0ac97b66-7233-4fcc-9b0f-856a67a62b4d" containerID="09101ceda6bf2ba07365f993acdcbbb18dd498f848daa46e32c04bb05065a77b" exitCode=0 Oct 01 09:53:36 crc kubenswrapper[4787]: I1001 09:53:36.910743 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-858fdb94fd-6tpq6" event={"ID":"0ac97b66-7233-4fcc-9b0f-856a67a62b4d","Type":"ContainerDied","Data":"09101ceda6bf2ba07365f993acdcbbb18dd498f848daa46e32c04bb05065a77b"} Oct 01 09:53:36 crc kubenswrapper[4787]: I1001 09:53:36.910882 4787 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 01 09:53:36 crc kubenswrapper[4787]: I1001 09:53:36.910892 4787 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 01 09:53:36 crc kubenswrapper[4787]: I1001 09:53:36.971304 4787 scope.go:117] "RemoveContainer" containerID="5784ec8cb42dd5d8922a1fcfb5e4e0b56515623a7de4cb21e67610f768fd5bc3" Oct 01 09:53:36 crc kubenswrapper[4787]: I1001 09:53:36.981541 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 09:53:36 crc kubenswrapper[4787]: I1001 09:53:36.993477 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.006451 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 09:53:37 crc kubenswrapper[4787]: E1001 09:53:37.006905 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6eff6348-9415-4d61-a405-0a6ac27869d4" containerName="probe" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.006918 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="6eff6348-9415-4d61-a405-0a6ac27869d4" containerName="probe" Oct 01 09:53:37 crc kubenswrapper[4787]: E1001 09:53:37.006949 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6eff6348-9415-4d61-a405-0a6ac27869d4" containerName="cinder-scheduler" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.006955 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="6eff6348-9415-4d61-a405-0a6ac27869d4" containerName="cinder-scheduler" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.007153 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="6eff6348-9415-4d61-a405-0a6ac27869d4" containerName="cinder-scheduler" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.007175 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="6eff6348-9415-4d61-a405-0a6ac27869d4" containerName="probe" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.010231 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.013201 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.016498 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.041909 4787 scope.go:117] "RemoveContainer" containerID="c9ecc450edb50582561268ad6d9c1a2bd4afd58b50ac192de75e9b62a534e973" Oct 01 09:53:37 crc kubenswrapper[4787]: E1001 09:53:37.046964 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9ecc450edb50582561268ad6d9c1a2bd4afd58b50ac192de75e9b62a534e973\": container with ID starting with c9ecc450edb50582561268ad6d9c1a2bd4afd58b50ac192de75e9b62a534e973 not found: ID does not exist" containerID="c9ecc450edb50582561268ad6d9c1a2bd4afd58b50ac192de75e9b62a534e973" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.047014 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9ecc450edb50582561268ad6d9c1a2bd4afd58b50ac192de75e9b62a534e973"} err="failed to get container status \"c9ecc450edb50582561268ad6d9c1a2bd4afd58b50ac192de75e9b62a534e973\": rpc error: code = NotFound desc = could not find container \"c9ecc450edb50582561268ad6d9c1a2bd4afd58b50ac192de75e9b62a534e973\": container with ID starting with c9ecc450edb50582561268ad6d9c1a2bd4afd58b50ac192de75e9b62a534e973 not found: ID does not exist" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.047042 4787 scope.go:117] "RemoveContainer" containerID="5784ec8cb42dd5d8922a1fcfb5e4e0b56515623a7de4cb21e67610f768fd5bc3" Oct 01 09:53:37 crc kubenswrapper[4787]: E1001 09:53:37.047439 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5784ec8cb42dd5d8922a1fcfb5e4e0b56515623a7de4cb21e67610f768fd5bc3\": container with ID starting with 5784ec8cb42dd5d8922a1fcfb5e4e0b56515623a7de4cb21e67610f768fd5bc3 not found: ID does not exist" containerID="5784ec8cb42dd5d8922a1fcfb5e4e0b56515623a7de4cb21e67610f768fd5bc3" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.047467 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5784ec8cb42dd5d8922a1fcfb5e4e0b56515623a7de4cb21e67610f768fd5bc3"} err="failed to get container status \"5784ec8cb42dd5d8922a1fcfb5e4e0b56515623a7de4cb21e67610f768fd5bc3\": rpc error: code = NotFound desc = could not find container \"5784ec8cb42dd5d8922a1fcfb5e4e0b56515623a7de4cb21e67610f768fd5bc3\": container with ID starting with 5784ec8cb42dd5d8922a1fcfb5e4e0b56515623a7de4cb21e67610f768fd5bc3 not found: ID does not exist" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.104668 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a844181-c2d3-4755-a0e2-d3706a9eb3e0-config-data\") pod \"cinder-scheduler-0\" (UID: \"8a844181-c2d3-4755-a0e2-d3706a9eb3e0\") " pod="openstack/cinder-scheduler-0" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.104734 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a844181-c2d3-4755-a0e2-d3706a9eb3e0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8a844181-c2d3-4755-a0e2-d3706a9eb3e0\") " pod="openstack/cinder-scheduler-0" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.104998 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8a844181-c2d3-4755-a0e2-d3706a9eb3e0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8a844181-c2d3-4755-a0e2-d3706a9eb3e0\") " pod="openstack/cinder-scheduler-0" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.105179 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8a844181-c2d3-4755-a0e2-d3706a9eb3e0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8a844181-c2d3-4755-a0e2-d3706a9eb3e0\") " pod="openstack/cinder-scheduler-0" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.105300 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a844181-c2d3-4755-a0e2-d3706a9eb3e0-scripts\") pod \"cinder-scheduler-0\" (UID: \"8a844181-c2d3-4755-a0e2-d3706a9eb3e0\") " pod="openstack/cinder-scheduler-0" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.105421 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxnzv\" (UniqueName: \"kubernetes.io/projected/8a844181-c2d3-4755-a0e2-d3706a9eb3e0-kube-api-access-mxnzv\") pod \"cinder-scheduler-0\" (UID: \"8a844181-c2d3-4755-a0e2-d3706a9eb3e0\") " pod="openstack/cinder-scheduler-0" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.186730 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-858fdb94fd-6tpq6" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.210958 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a844181-c2d3-4755-a0e2-d3706a9eb3e0-config-data\") pod \"cinder-scheduler-0\" (UID: \"8a844181-c2d3-4755-a0e2-d3706a9eb3e0\") " pod="openstack/cinder-scheduler-0" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.211030 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a844181-c2d3-4755-a0e2-d3706a9eb3e0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8a844181-c2d3-4755-a0e2-d3706a9eb3e0\") " pod="openstack/cinder-scheduler-0" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.211264 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8a844181-c2d3-4755-a0e2-d3706a9eb3e0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8a844181-c2d3-4755-a0e2-d3706a9eb3e0\") " pod="openstack/cinder-scheduler-0" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.211306 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8a844181-c2d3-4755-a0e2-d3706a9eb3e0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8a844181-c2d3-4755-a0e2-d3706a9eb3e0\") " pod="openstack/cinder-scheduler-0" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.211348 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a844181-c2d3-4755-a0e2-d3706a9eb3e0-scripts\") pod \"cinder-scheduler-0\" (UID: \"8a844181-c2d3-4755-a0e2-d3706a9eb3e0\") " pod="openstack/cinder-scheduler-0" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.211387 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxnzv\" (UniqueName: \"kubernetes.io/projected/8a844181-c2d3-4755-a0e2-d3706a9eb3e0-kube-api-access-mxnzv\") pod \"cinder-scheduler-0\" (UID: \"8a844181-c2d3-4755-a0e2-d3706a9eb3e0\") " pod="openstack/cinder-scheduler-0" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.220262 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8a844181-c2d3-4755-a0e2-d3706a9eb3e0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8a844181-c2d3-4755-a0e2-d3706a9eb3e0\") " pod="openstack/cinder-scheduler-0" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.224633 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a844181-c2d3-4755-a0e2-d3706a9eb3e0-config-data\") pod \"cinder-scheduler-0\" (UID: \"8a844181-c2d3-4755-a0e2-d3706a9eb3e0\") " pod="openstack/cinder-scheduler-0" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.224632 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a844181-c2d3-4755-a0e2-d3706a9eb3e0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8a844181-c2d3-4755-a0e2-d3706a9eb3e0\") " pod="openstack/cinder-scheduler-0" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.246881 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxnzv\" (UniqueName: \"kubernetes.io/projected/8a844181-c2d3-4755-a0e2-d3706a9eb3e0-kube-api-access-mxnzv\") pod \"cinder-scheduler-0\" (UID: \"8a844181-c2d3-4755-a0e2-d3706a9eb3e0\") " pod="openstack/cinder-scheduler-0" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.248790 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a844181-c2d3-4755-a0e2-d3706a9eb3e0-scripts\") pod \"cinder-scheduler-0\" (UID: \"8a844181-c2d3-4755-a0e2-d3706a9eb3e0\") " pod="openstack/cinder-scheduler-0" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.252697 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8a844181-c2d3-4755-a0e2-d3706a9eb3e0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8a844181-c2d3-4755-a0e2-d3706a9eb3e0\") " pod="openstack/cinder-scheduler-0" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.314148 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lq7kt\" (UniqueName: \"kubernetes.io/projected/0ac97b66-7233-4fcc-9b0f-856a67a62b4d-kube-api-access-lq7kt\") pod \"0ac97b66-7233-4fcc-9b0f-856a67a62b4d\" (UID: \"0ac97b66-7233-4fcc-9b0f-856a67a62b4d\") " Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.314360 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0ac97b66-7233-4fcc-9b0f-856a67a62b4d-httpd-config\") pod \"0ac97b66-7233-4fcc-9b0f-856a67a62b4d\" (UID: \"0ac97b66-7233-4fcc-9b0f-856a67a62b4d\") " Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.314387 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ac97b66-7233-4fcc-9b0f-856a67a62b4d-ovndb-tls-certs\") pod \"0ac97b66-7233-4fcc-9b0f-856a67a62b4d\" (UID: \"0ac97b66-7233-4fcc-9b0f-856a67a62b4d\") " Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.314491 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0ac97b66-7233-4fcc-9b0f-856a67a62b4d-config\") pod \"0ac97b66-7233-4fcc-9b0f-856a67a62b4d\" (UID: \"0ac97b66-7233-4fcc-9b0f-856a67a62b4d\") " Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.314575 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ac97b66-7233-4fcc-9b0f-856a67a62b4d-combined-ca-bundle\") pod \"0ac97b66-7233-4fcc-9b0f-856a67a62b4d\" (UID: \"0ac97b66-7233-4fcc-9b0f-856a67a62b4d\") " Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.317901 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ac97b66-7233-4fcc-9b0f-856a67a62b4d-kube-api-access-lq7kt" (OuterVolumeSpecName: "kube-api-access-lq7kt") pod "0ac97b66-7233-4fcc-9b0f-856a67a62b4d" (UID: "0ac97b66-7233-4fcc-9b0f-856a67a62b4d"). InnerVolumeSpecName "kube-api-access-lq7kt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.322635 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ac97b66-7233-4fcc-9b0f-856a67a62b4d-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "0ac97b66-7233-4fcc-9b0f-856a67a62b4d" (UID: "0ac97b66-7233-4fcc-9b0f-856a67a62b4d"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.342166 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.393110 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ac97b66-7233-4fcc-9b0f-856a67a62b4d-config" (OuterVolumeSpecName: "config") pod "0ac97b66-7233-4fcc-9b0f-856a67a62b4d" (UID: "0ac97b66-7233-4fcc-9b0f-856a67a62b4d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.418016 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lq7kt\" (UniqueName: \"kubernetes.io/projected/0ac97b66-7233-4fcc-9b0f-856a67a62b4d-kube-api-access-lq7kt\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.418059 4787 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0ac97b66-7233-4fcc-9b0f-856a67a62b4d-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.418090 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/0ac97b66-7233-4fcc-9b0f-856a67a62b4d-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.426933 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ac97b66-7233-4fcc-9b0f-856a67a62b4d-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "0ac97b66-7233-4fcc-9b0f-856a67a62b4d" (UID: "0ac97b66-7233-4fcc-9b0f-856a67a62b4d"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.440180 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ac97b66-7233-4fcc-9b0f-856a67a62b4d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0ac97b66-7233-4fcc-9b0f-856a67a62b4d" (UID: "0ac97b66-7233-4fcc-9b0f-856a67a62b4d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.520174 4787 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ac97b66-7233-4fcc-9b0f-856a67a62b4d-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.520493 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ac97b66-7233-4fcc-9b0f-856a67a62b4d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.562245 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.564421 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.927313 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-858fdb94fd-6tpq6" event={"ID":"0ac97b66-7233-4fcc-9b0f-856a67a62b4d","Type":"ContainerDied","Data":"879a9663415d32abcb590d620ef91ad608f4b823f66c8ca7a06051798ee1e67d"} Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.927707 4787 scope.go:117] "RemoveContainer" containerID="2b5867e80774c39c9e0b986b908a0d57bfbab946cc4e15e39311fc5003ff06a0" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.927596 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-858fdb94fd-6tpq6" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.930143 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.936839 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14c412a3-bcdc-4a19-b769-310820125df0","Type":"ContainerStarted","Data":"949b5aca8fd3ff42ce062ecb5c220a87429fed548318c3024c3a86168f30f764"} Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.938265 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-9192-account-create-rcrfw"] Oct 01 09:53:37 crc kubenswrapper[4787]: E1001 09:53:37.938899 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ac97b66-7233-4fcc-9b0f-856a67a62b4d" containerName="neutron-httpd" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.938925 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ac97b66-7233-4fcc-9b0f-856a67a62b4d" containerName="neutron-httpd" Oct 01 09:53:37 crc kubenswrapper[4787]: E1001 09:53:37.939005 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ac97b66-7233-4fcc-9b0f-856a67a62b4d" containerName="neutron-api" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.939017 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ac97b66-7233-4fcc-9b0f-856a67a62b4d" containerName="neutron-api" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.939299 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ac97b66-7233-4fcc-9b0f-856a67a62b4d" containerName="neutron-api" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.939330 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ac97b66-7233-4fcc-9b0f-856a67a62b4d" containerName="neutron-httpd" Oct 01 09:53:37 crc kubenswrapper[4787]: W1001 09:53:37.942181 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a844181_c2d3_4755_a0e2_d3706a9eb3e0.slice/crio-eba0caff4db4556b744e57f712fd4272807944ef691bc114ffa389c575c65d16 WatchSource:0}: Error finding container eba0caff4db4556b744e57f712fd4272807944ef691bc114ffa389c575c65d16: Status 404 returned error can't find the container with id eba0caff4db4556b744e57f712fd4272807944ef691bc114ffa389c575c65d16 Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.947380 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-9192-account-create-rcrfw"] Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.947520 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9192-account-create-rcrfw" Oct 01 09:53:37 crc kubenswrapper[4787]: I1001 09:53:37.951385 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 01 09:53:38 crc kubenswrapper[4787]: I1001 09:53:38.039802 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvqsk\" (UniqueName: \"kubernetes.io/projected/77912320-f656-463d-a030-551ff526f530-kube-api-access-zvqsk\") pod \"nova-cell0-9192-account-create-rcrfw\" (UID: \"77912320-f656-463d-a030-551ff526f530\") " pod="openstack/nova-cell0-9192-account-create-rcrfw" Oct 01 09:53:38 crc kubenswrapper[4787]: I1001 09:53:38.082962 4787 scope.go:117] "RemoveContainer" containerID="09101ceda6bf2ba07365f993acdcbbb18dd498f848daa46e32c04bb05065a77b" Oct 01 09:53:38 crc kubenswrapper[4787]: I1001 09:53:38.084689 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-858fdb94fd-6tpq6"] Oct 01 09:53:38 crc kubenswrapper[4787]: I1001 09:53:38.094553 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-858fdb94fd-6tpq6"] Oct 01 09:53:38 crc kubenswrapper[4787]: I1001 09:53:38.123844 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-9f21-account-create-8zqzm"] Oct 01 09:53:38 crc kubenswrapper[4787]: I1001 09:53:38.124960 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-9f21-account-create-8zqzm" Oct 01 09:53:38 crc kubenswrapper[4787]: I1001 09:53:38.129981 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 01 09:53:38 crc kubenswrapper[4787]: I1001 09:53:38.135685 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-9f21-account-create-8zqzm"] Oct 01 09:53:38 crc kubenswrapper[4787]: I1001 09:53:38.143035 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvqsk\" (UniqueName: \"kubernetes.io/projected/77912320-f656-463d-a030-551ff526f530-kube-api-access-zvqsk\") pod \"nova-cell0-9192-account-create-rcrfw\" (UID: \"77912320-f656-463d-a030-551ff526f530\") " pod="openstack/nova-cell0-9192-account-create-rcrfw" Oct 01 09:53:38 crc kubenswrapper[4787]: I1001 09:53:38.159873 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvqsk\" (UniqueName: \"kubernetes.io/projected/77912320-f656-463d-a030-551ff526f530-kube-api-access-zvqsk\") pod \"nova-cell0-9192-account-create-rcrfw\" (UID: \"77912320-f656-463d-a030-551ff526f530\") " pod="openstack/nova-cell0-9192-account-create-rcrfw" Oct 01 09:53:38 crc kubenswrapper[4787]: I1001 09:53:38.245663 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlqqn\" (UniqueName: \"kubernetes.io/projected/00ce9c3d-e2af-4c96-a018-87539fbe9b42-kube-api-access-xlqqn\") pod \"nova-cell1-9f21-account-create-8zqzm\" (UID: \"00ce9c3d-e2af-4c96-a018-87539fbe9b42\") " pod="openstack/nova-cell1-9f21-account-create-8zqzm" Oct 01 09:53:38 crc kubenswrapper[4787]: I1001 09:53:38.347707 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlqqn\" (UniqueName: \"kubernetes.io/projected/00ce9c3d-e2af-4c96-a018-87539fbe9b42-kube-api-access-xlqqn\") pod \"nova-cell1-9f21-account-create-8zqzm\" (UID: \"00ce9c3d-e2af-4c96-a018-87539fbe9b42\") " pod="openstack/nova-cell1-9f21-account-create-8zqzm" Oct 01 09:53:38 crc kubenswrapper[4787]: I1001 09:53:38.373530 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9192-account-create-rcrfw" Oct 01 09:53:38 crc kubenswrapper[4787]: I1001 09:53:38.374892 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlqqn\" (UniqueName: \"kubernetes.io/projected/00ce9c3d-e2af-4c96-a018-87539fbe9b42-kube-api-access-xlqqn\") pod \"nova-cell1-9f21-account-create-8zqzm\" (UID: \"00ce9c3d-e2af-4c96-a018-87539fbe9b42\") " pod="openstack/nova-cell1-9f21-account-create-8zqzm" Oct 01 09:53:38 crc kubenswrapper[4787]: I1001 09:53:38.446227 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-9f21-account-create-8zqzm" Oct 01 09:53:38 crc kubenswrapper[4787]: I1001 09:53:38.544018 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ac97b66-7233-4fcc-9b0f-856a67a62b4d" path="/var/lib/kubelet/pods/0ac97b66-7233-4fcc-9b0f-856a67a62b4d/volumes" Oct 01 09:53:38 crc kubenswrapper[4787]: I1001 09:53:38.545424 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6eff6348-9415-4d61-a405-0a6ac27869d4" path="/var/lib/kubelet/pods/6eff6348-9415-4d61-a405-0a6ac27869d4/volumes" Oct 01 09:53:38 crc kubenswrapper[4787]: I1001 09:53:38.967221 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8a844181-c2d3-4755-a0e2-d3706a9eb3e0","Type":"ContainerStarted","Data":"eba0caff4db4556b744e57f712fd4272807944ef691bc114ffa389c575c65d16"} Oct 01 09:53:39 crc kubenswrapper[4787]: I1001 09:53:39.011256 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-9192-account-create-rcrfw"] Oct 01 09:53:39 crc kubenswrapper[4787]: I1001 09:53:39.126542 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-9f21-account-create-8zqzm"] Oct 01 09:53:39 crc kubenswrapper[4787]: W1001 09:53:39.127649 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00ce9c3d_e2af_4c96_a018_87539fbe9b42.slice/crio-df50fce21893102babade4b3b0b272611fe1885c1c5c99737e153c4b019ff568 WatchSource:0}: Error finding container df50fce21893102babade4b3b0b272611fe1885c1c5c99737e153c4b019ff568: Status 404 returned error can't find the container with id df50fce21893102babade4b3b0b272611fe1885c1c5c99737e153c4b019ff568 Oct 01 09:53:39 crc kubenswrapper[4787]: I1001 09:53:39.982979 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8a844181-c2d3-4755-a0e2-d3706a9eb3e0","Type":"ContainerStarted","Data":"5778e6872116cfa240ef058eac44ecae89109f97be3dd5dfe3c67c9c314262b2"} Oct 01 09:53:39 crc kubenswrapper[4787]: I1001 09:53:39.986501 4787 generic.go:334] "Generic (PLEG): container finished" podID="77912320-f656-463d-a030-551ff526f530" containerID="431a329239e735143b5e81c9f3853dc3171c5a42ceda423213bbd0424a81afa7" exitCode=0 Oct 01 09:53:39 crc kubenswrapper[4787]: I1001 09:53:39.986565 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9192-account-create-rcrfw" event={"ID":"77912320-f656-463d-a030-551ff526f530","Type":"ContainerDied","Data":"431a329239e735143b5e81c9f3853dc3171c5a42ceda423213bbd0424a81afa7"} Oct 01 09:53:39 crc kubenswrapper[4787]: I1001 09:53:39.986591 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9192-account-create-rcrfw" event={"ID":"77912320-f656-463d-a030-551ff526f530","Type":"ContainerStarted","Data":"29c30ff35dd2bd444879eeb3bea240f0df6f7e5e4ed8a8e9b1597c236c3aadec"} Oct 01 09:53:39 crc kubenswrapper[4787]: I1001 09:53:39.990486 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14c412a3-bcdc-4a19-b769-310820125df0","Type":"ContainerStarted","Data":"259adfed6bd06e27e49c4f41edf0c5edb0070ac71bdea39e66a2781d171c6aee"} Oct 01 09:53:39 crc kubenswrapper[4787]: I1001 09:53:39.990631 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="14c412a3-bcdc-4a19-b769-310820125df0" containerName="ceilometer-central-agent" containerID="cri-o://9fef4997b8b7315a153484f251256d79d41f4ad455d3e9db1fba2a85071cfd46" gracePeriod=30 Oct 01 09:53:39 crc kubenswrapper[4787]: I1001 09:53:39.990706 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 01 09:53:39 crc kubenswrapper[4787]: I1001 09:53:39.990746 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="14c412a3-bcdc-4a19-b769-310820125df0" containerName="proxy-httpd" containerID="cri-o://259adfed6bd06e27e49c4f41edf0c5edb0070ac71bdea39e66a2781d171c6aee" gracePeriod=30 Oct 01 09:53:39 crc kubenswrapper[4787]: I1001 09:53:39.990785 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="14c412a3-bcdc-4a19-b769-310820125df0" containerName="sg-core" containerID="cri-o://949b5aca8fd3ff42ce062ecb5c220a87429fed548318c3024c3a86168f30f764" gracePeriod=30 Oct 01 09:53:39 crc kubenswrapper[4787]: I1001 09:53:39.990814 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="14c412a3-bcdc-4a19-b769-310820125df0" containerName="ceilometer-notification-agent" containerID="cri-o://820b434e99203b0219ec476dc6a6a9fdfe43a8a8957692874fe905ef8dc37585" gracePeriod=30 Oct 01 09:53:39 crc kubenswrapper[4787]: I1001 09:53:39.994556 4787 generic.go:334] "Generic (PLEG): container finished" podID="00ce9c3d-e2af-4c96-a018-87539fbe9b42" containerID="a4ea6073b367bb6bb5a6b790a4f7e3b223a52f0e0035eef4de083e15ebddf3b8" exitCode=0 Oct 01 09:53:39 crc kubenswrapper[4787]: I1001 09:53:39.994598 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-9f21-account-create-8zqzm" event={"ID":"00ce9c3d-e2af-4c96-a018-87539fbe9b42","Type":"ContainerDied","Data":"a4ea6073b367bb6bb5a6b790a4f7e3b223a52f0e0035eef4de083e15ebddf3b8"} Oct 01 09:53:39 crc kubenswrapper[4787]: I1001 09:53:39.994622 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-9f21-account-create-8zqzm" event={"ID":"00ce9c3d-e2af-4c96-a018-87539fbe9b42","Type":"ContainerStarted","Data":"df50fce21893102babade4b3b0b272611fe1885c1c5c99737e153c4b019ff568"} Oct 01 09:53:40 crc kubenswrapper[4787]: I1001 09:53:40.071278 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.089950909 podStartE2EDuration="7.07125619s" podCreationTimestamp="2025-10-01 09:53:33 +0000 UTC" firstStartedPulling="2025-10-01 09:53:34.797856793 +0000 UTC m=+1046.913000950" lastFinishedPulling="2025-10-01 09:53:38.779162074 +0000 UTC m=+1050.894306231" observedRunningTime="2025-10-01 09:53:40.064864228 +0000 UTC m=+1052.180008385" watchObservedRunningTime="2025-10-01 09:53:40.07125619 +0000 UTC m=+1052.186400347" Oct 01 09:53:41 crc kubenswrapper[4787]: I1001 09:53:41.007455 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8a844181-c2d3-4755-a0e2-d3706a9eb3e0","Type":"ContainerStarted","Data":"58829b2fb6cc6c40100f17a30ff9e8c18aba6ea86d56a4267c76d38206459dd8"} Oct 01 09:53:41 crc kubenswrapper[4787]: I1001 09:53:41.011026 4787 generic.go:334] "Generic (PLEG): container finished" podID="14c412a3-bcdc-4a19-b769-310820125df0" containerID="259adfed6bd06e27e49c4f41edf0c5edb0070ac71bdea39e66a2781d171c6aee" exitCode=0 Oct 01 09:53:41 crc kubenswrapper[4787]: I1001 09:53:41.011050 4787 generic.go:334] "Generic (PLEG): container finished" podID="14c412a3-bcdc-4a19-b769-310820125df0" containerID="949b5aca8fd3ff42ce062ecb5c220a87429fed548318c3024c3a86168f30f764" exitCode=2 Oct 01 09:53:41 crc kubenswrapper[4787]: I1001 09:53:41.011057 4787 generic.go:334] "Generic (PLEG): container finished" podID="14c412a3-bcdc-4a19-b769-310820125df0" containerID="820b434e99203b0219ec476dc6a6a9fdfe43a8a8957692874fe905ef8dc37585" exitCode=0 Oct 01 09:53:41 crc kubenswrapper[4787]: I1001 09:53:41.011099 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14c412a3-bcdc-4a19-b769-310820125df0","Type":"ContainerDied","Data":"259adfed6bd06e27e49c4f41edf0c5edb0070ac71bdea39e66a2781d171c6aee"} Oct 01 09:53:41 crc kubenswrapper[4787]: I1001 09:53:41.011125 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14c412a3-bcdc-4a19-b769-310820125df0","Type":"ContainerDied","Data":"949b5aca8fd3ff42ce062ecb5c220a87429fed548318c3024c3a86168f30f764"} Oct 01 09:53:41 crc kubenswrapper[4787]: I1001 09:53:41.011134 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14c412a3-bcdc-4a19-b769-310820125df0","Type":"ContainerDied","Data":"820b434e99203b0219ec476dc6a6a9fdfe43a8a8957692874fe905ef8dc37585"} Oct 01 09:53:41 crc kubenswrapper[4787]: I1001 09:53:41.038066 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.03804307 podStartE2EDuration="5.03804307s" podCreationTimestamp="2025-10-01 09:53:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:53:41.033660669 +0000 UTC m=+1053.148804836" watchObservedRunningTime="2025-10-01 09:53:41.03804307 +0000 UTC m=+1053.153187237" Oct 01 09:53:41 crc kubenswrapper[4787]: I1001 09:53:41.251120 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:53:41 crc kubenswrapper[4787]: I1001 09:53:41.251511 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:53:41 crc kubenswrapper[4787]: I1001 09:53:41.555902 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-9f21-account-create-8zqzm" Oct 01 09:53:41 crc kubenswrapper[4787]: I1001 09:53:41.563569 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9192-account-create-rcrfw" Oct 01 09:53:41 crc kubenswrapper[4787]: I1001 09:53:41.644397 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvqsk\" (UniqueName: \"kubernetes.io/projected/77912320-f656-463d-a030-551ff526f530-kube-api-access-zvqsk\") pod \"77912320-f656-463d-a030-551ff526f530\" (UID: \"77912320-f656-463d-a030-551ff526f530\") " Oct 01 09:53:41 crc kubenswrapper[4787]: I1001 09:53:41.644466 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xlqqn\" (UniqueName: \"kubernetes.io/projected/00ce9c3d-e2af-4c96-a018-87539fbe9b42-kube-api-access-xlqqn\") pod \"00ce9c3d-e2af-4c96-a018-87539fbe9b42\" (UID: \"00ce9c3d-e2af-4c96-a018-87539fbe9b42\") " Oct 01 09:53:41 crc kubenswrapper[4787]: I1001 09:53:41.654033 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00ce9c3d-e2af-4c96-a018-87539fbe9b42-kube-api-access-xlqqn" (OuterVolumeSpecName: "kube-api-access-xlqqn") pod "00ce9c3d-e2af-4c96-a018-87539fbe9b42" (UID: "00ce9c3d-e2af-4c96-a018-87539fbe9b42"). InnerVolumeSpecName "kube-api-access-xlqqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:53:41 crc kubenswrapper[4787]: I1001 09:53:41.656166 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77912320-f656-463d-a030-551ff526f530-kube-api-access-zvqsk" (OuterVolumeSpecName: "kube-api-access-zvqsk") pod "77912320-f656-463d-a030-551ff526f530" (UID: "77912320-f656-463d-a030-551ff526f530"). InnerVolumeSpecName "kube-api-access-zvqsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:53:41 crc kubenswrapper[4787]: I1001 09:53:41.748804 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvqsk\" (UniqueName: \"kubernetes.io/projected/77912320-f656-463d-a030-551ff526f530-kube-api-access-zvqsk\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:41 crc kubenswrapper[4787]: I1001 09:53:41.748843 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xlqqn\" (UniqueName: \"kubernetes.io/projected/00ce9c3d-e2af-4c96-a018-87539fbe9b42-kube-api-access-xlqqn\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:42 crc kubenswrapper[4787]: I1001 09:53:42.038537 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-9f21-account-create-8zqzm" Oct 01 09:53:42 crc kubenswrapper[4787]: I1001 09:53:42.039341 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-9f21-account-create-8zqzm" event={"ID":"00ce9c3d-e2af-4c96-a018-87539fbe9b42","Type":"ContainerDied","Data":"df50fce21893102babade4b3b0b272611fe1885c1c5c99737e153c4b019ff568"} Oct 01 09:53:42 crc kubenswrapper[4787]: I1001 09:53:42.039392 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df50fce21893102babade4b3b0b272611fe1885c1c5c99737e153c4b019ff568" Oct 01 09:53:42 crc kubenswrapper[4787]: I1001 09:53:42.048168 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9192-account-create-rcrfw" event={"ID":"77912320-f656-463d-a030-551ff526f530","Type":"ContainerDied","Data":"29c30ff35dd2bd444879eeb3bea240f0df6f7e5e4ed8a8e9b1597c236c3aadec"} Oct 01 09:53:42 crc kubenswrapper[4787]: I1001 09:53:42.048226 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="29c30ff35dd2bd444879eeb3bea240f0df6f7e5e4ed8a8e9b1597c236c3aadec" Oct 01 09:53:42 crc kubenswrapper[4787]: I1001 09:53:42.048223 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9192-account-create-rcrfw" Oct 01 09:53:42 crc kubenswrapper[4787]: I1001 09:53:42.242879 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 01 09:53:42 crc kubenswrapper[4787]: I1001 09:53:42.346195 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 01 09:53:43 crc kubenswrapper[4787]: I1001 09:53:43.276049 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-tj64b"] Oct 01 09:53:43 crc kubenswrapper[4787]: E1001 09:53:43.277147 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00ce9c3d-e2af-4c96-a018-87539fbe9b42" containerName="mariadb-account-create" Oct 01 09:53:43 crc kubenswrapper[4787]: I1001 09:53:43.277163 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="00ce9c3d-e2af-4c96-a018-87539fbe9b42" containerName="mariadb-account-create" Oct 01 09:53:43 crc kubenswrapper[4787]: E1001 09:53:43.277187 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77912320-f656-463d-a030-551ff526f530" containerName="mariadb-account-create" Oct 01 09:53:43 crc kubenswrapper[4787]: I1001 09:53:43.277193 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="77912320-f656-463d-a030-551ff526f530" containerName="mariadb-account-create" Oct 01 09:53:43 crc kubenswrapper[4787]: I1001 09:53:43.277394 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="00ce9c3d-e2af-4c96-a018-87539fbe9b42" containerName="mariadb-account-create" Oct 01 09:53:43 crc kubenswrapper[4787]: I1001 09:53:43.277404 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="77912320-f656-463d-a030-551ff526f530" containerName="mariadb-account-create" Oct 01 09:53:43 crc kubenswrapper[4787]: I1001 09:53:43.278021 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-tj64b" Oct 01 09:53:43 crc kubenswrapper[4787]: I1001 09:53:43.280597 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 01 09:53:43 crc kubenswrapper[4787]: I1001 09:53:43.280756 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 01 09:53:43 crc kubenswrapper[4787]: I1001 09:53:43.280932 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-hjcxb" Oct 01 09:53:43 crc kubenswrapper[4787]: I1001 09:53:43.290797 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-tj64b"] Oct 01 09:53:43 crc kubenswrapper[4787]: I1001 09:53:43.381465 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9821f091-0513-4370-beb7-8239eab74972-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-tj64b\" (UID: \"9821f091-0513-4370-beb7-8239eab74972\") " pod="openstack/nova-cell0-conductor-db-sync-tj64b" Oct 01 09:53:43 crc kubenswrapper[4787]: I1001 09:53:43.381538 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9821f091-0513-4370-beb7-8239eab74972-scripts\") pod \"nova-cell0-conductor-db-sync-tj64b\" (UID: \"9821f091-0513-4370-beb7-8239eab74972\") " pod="openstack/nova-cell0-conductor-db-sync-tj64b" Oct 01 09:53:43 crc kubenswrapper[4787]: I1001 09:53:43.381655 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhbl5\" (UniqueName: \"kubernetes.io/projected/9821f091-0513-4370-beb7-8239eab74972-kube-api-access-lhbl5\") pod \"nova-cell0-conductor-db-sync-tj64b\" (UID: \"9821f091-0513-4370-beb7-8239eab74972\") " pod="openstack/nova-cell0-conductor-db-sync-tj64b" Oct 01 09:53:43 crc kubenswrapper[4787]: I1001 09:53:43.381682 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9821f091-0513-4370-beb7-8239eab74972-config-data\") pod \"nova-cell0-conductor-db-sync-tj64b\" (UID: \"9821f091-0513-4370-beb7-8239eab74972\") " pod="openstack/nova-cell0-conductor-db-sync-tj64b" Oct 01 09:53:43 crc kubenswrapper[4787]: I1001 09:53:43.483480 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9821f091-0513-4370-beb7-8239eab74972-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-tj64b\" (UID: \"9821f091-0513-4370-beb7-8239eab74972\") " pod="openstack/nova-cell0-conductor-db-sync-tj64b" Oct 01 09:53:43 crc kubenswrapper[4787]: I1001 09:53:43.483562 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9821f091-0513-4370-beb7-8239eab74972-scripts\") pod \"nova-cell0-conductor-db-sync-tj64b\" (UID: \"9821f091-0513-4370-beb7-8239eab74972\") " pod="openstack/nova-cell0-conductor-db-sync-tj64b" Oct 01 09:53:43 crc kubenswrapper[4787]: I1001 09:53:43.483610 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhbl5\" (UniqueName: \"kubernetes.io/projected/9821f091-0513-4370-beb7-8239eab74972-kube-api-access-lhbl5\") pod \"nova-cell0-conductor-db-sync-tj64b\" (UID: \"9821f091-0513-4370-beb7-8239eab74972\") " pod="openstack/nova-cell0-conductor-db-sync-tj64b" Oct 01 09:53:43 crc kubenswrapper[4787]: I1001 09:53:43.483642 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9821f091-0513-4370-beb7-8239eab74972-config-data\") pod \"nova-cell0-conductor-db-sync-tj64b\" (UID: \"9821f091-0513-4370-beb7-8239eab74972\") " pod="openstack/nova-cell0-conductor-db-sync-tj64b" Oct 01 09:53:43 crc kubenswrapper[4787]: I1001 09:53:43.489730 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9821f091-0513-4370-beb7-8239eab74972-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-tj64b\" (UID: \"9821f091-0513-4370-beb7-8239eab74972\") " pod="openstack/nova-cell0-conductor-db-sync-tj64b" Oct 01 09:53:43 crc kubenswrapper[4787]: I1001 09:53:43.490229 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9821f091-0513-4370-beb7-8239eab74972-config-data\") pod \"nova-cell0-conductor-db-sync-tj64b\" (UID: \"9821f091-0513-4370-beb7-8239eab74972\") " pod="openstack/nova-cell0-conductor-db-sync-tj64b" Oct 01 09:53:43 crc kubenswrapper[4787]: I1001 09:53:43.491657 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9821f091-0513-4370-beb7-8239eab74972-scripts\") pod \"nova-cell0-conductor-db-sync-tj64b\" (UID: \"9821f091-0513-4370-beb7-8239eab74972\") " pod="openstack/nova-cell0-conductor-db-sync-tj64b" Oct 01 09:53:43 crc kubenswrapper[4787]: I1001 09:53:43.502301 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhbl5\" (UniqueName: \"kubernetes.io/projected/9821f091-0513-4370-beb7-8239eab74972-kube-api-access-lhbl5\") pod \"nova-cell0-conductor-db-sync-tj64b\" (UID: \"9821f091-0513-4370-beb7-8239eab74972\") " pod="openstack/nova-cell0-conductor-db-sync-tj64b" Oct 01 09:53:43 crc kubenswrapper[4787]: I1001 09:53:43.596589 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-tj64b" Oct 01 09:53:44 crc kubenswrapper[4787]: I1001 09:53:44.175099 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-tj64b"] Oct 01 09:53:45 crc kubenswrapper[4787]: I1001 09:53:45.082698 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-tj64b" event={"ID":"9821f091-0513-4370-beb7-8239eab74972","Type":"ContainerStarted","Data":"cabafafa3aa175c401aef7fdcc6f183dfaba812c9865c6c1ba582206ef9f5688"} Oct 01 09:53:45 crc kubenswrapper[4787]: I1001 09:53:45.735468 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:53:45 crc kubenswrapper[4787]: I1001 09:53:45.834103 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14c412a3-bcdc-4a19-b769-310820125df0-combined-ca-bundle\") pod \"14c412a3-bcdc-4a19-b769-310820125df0\" (UID: \"14c412a3-bcdc-4a19-b769-310820125df0\") " Oct 01 09:53:45 crc kubenswrapper[4787]: I1001 09:53:45.834234 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14c412a3-bcdc-4a19-b769-310820125df0-log-httpd\") pod \"14c412a3-bcdc-4a19-b769-310820125df0\" (UID: \"14c412a3-bcdc-4a19-b769-310820125df0\") " Oct 01 09:53:45 crc kubenswrapper[4787]: I1001 09:53:45.834269 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14c412a3-bcdc-4a19-b769-310820125df0-scripts\") pod \"14c412a3-bcdc-4a19-b769-310820125df0\" (UID: \"14c412a3-bcdc-4a19-b769-310820125df0\") " Oct 01 09:53:45 crc kubenswrapper[4787]: I1001 09:53:45.834340 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14c412a3-bcdc-4a19-b769-310820125df0-run-httpd\") pod \"14c412a3-bcdc-4a19-b769-310820125df0\" (UID: \"14c412a3-bcdc-4a19-b769-310820125df0\") " Oct 01 09:53:45 crc kubenswrapper[4787]: I1001 09:53:45.834418 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14c412a3-bcdc-4a19-b769-310820125df0-config-data\") pod \"14c412a3-bcdc-4a19-b769-310820125df0\" (UID: \"14c412a3-bcdc-4a19-b769-310820125df0\") " Oct 01 09:53:45 crc kubenswrapper[4787]: I1001 09:53:45.834436 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/14c412a3-bcdc-4a19-b769-310820125df0-sg-core-conf-yaml\") pod \"14c412a3-bcdc-4a19-b769-310820125df0\" (UID: \"14c412a3-bcdc-4a19-b769-310820125df0\") " Oct 01 09:53:45 crc kubenswrapper[4787]: I1001 09:53:45.834454 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zbxtt\" (UniqueName: \"kubernetes.io/projected/14c412a3-bcdc-4a19-b769-310820125df0-kube-api-access-zbxtt\") pod \"14c412a3-bcdc-4a19-b769-310820125df0\" (UID: \"14c412a3-bcdc-4a19-b769-310820125df0\") " Oct 01 09:53:45 crc kubenswrapper[4787]: I1001 09:53:45.835022 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14c412a3-bcdc-4a19-b769-310820125df0-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "14c412a3-bcdc-4a19-b769-310820125df0" (UID: "14c412a3-bcdc-4a19-b769-310820125df0"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:53:45 crc kubenswrapper[4787]: I1001 09:53:45.835225 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14c412a3-bcdc-4a19-b769-310820125df0-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "14c412a3-bcdc-4a19-b769-310820125df0" (UID: "14c412a3-bcdc-4a19-b769-310820125df0"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:53:45 crc kubenswrapper[4787]: I1001 09:53:45.840196 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14c412a3-bcdc-4a19-b769-310820125df0-scripts" (OuterVolumeSpecName: "scripts") pod "14c412a3-bcdc-4a19-b769-310820125df0" (UID: "14c412a3-bcdc-4a19-b769-310820125df0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:45 crc kubenswrapper[4787]: I1001 09:53:45.840360 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14c412a3-bcdc-4a19-b769-310820125df0-kube-api-access-zbxtt" (OuterVolumeSpecName: "kube-api-access-zbxtt") pod "14c412a3-bcdc-4a19-b769-310820125df0" (UID: "14c412a3-bcdc-4a19-b769-310820125df0"). InnerVolumeSpecName "kube-api-access-zbxtt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:53:45 crc kubenswrapper[4787]: I1001 09:53:45.876228 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14c412a3-bcdc-4a19-b769-310820125df0-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "14c412a3-bcdc-4a19-b769-310820125df0" (UID: "14c412a3-bcdc-4a19-b769-310820125df0"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:45 crc kubenswrapper[4787]: I1001 09:53:45.918867 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14c412a3-bcdc-4a19-b769-310820125df0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "14c412a3-bcdc-4a19-b769-310820125df0" (UID: "14c412a3-bcdc-4a19-b769-310820125df0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:45 crc kubenswrapper[4787]: I1001 09:53:45.936532 4787 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/14c412a3-bcdc-4a19-b769-310820125df0-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:45 crc kubenswrapper[4787]: I1001 09:53:45.936579 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zbxtt\" (UniqueName: \"kubernetes.io/projected/14c412a3-bcdc-4a19-b769-310820125df0-kube-api-access-zbxtt\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:45 crc kubenswrapper[4787]: I1001 09:53:45.936590 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14c412a3-bcdc-4a19-b769-310820125df0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:45 crc kubenswrapper[4787]: I1001 09:53:45.936600 4787 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14c412a3-bcdc-4a19-b769-310820125df0-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:45 crc kubenswrapper[4787]: I1001 09:53:45.936609 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14c412a3-bcdc-4a19-b769-310820125df0-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:45 crc kubenswrapper[4787]: I1001 09:53:45.936618 4787 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14c412a3-bcdc-4a19-b769-310820125df0-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:45 crc kubenswrapper[4787]: I1001 09:53:45.938723 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14c412a3-bcdc-4a19-b769-310820125df0-config-data" (OuterVolumeSpecName: "config-data") pod "14c412a3-bcdc-4a19-b769-310820125df0" (UID: "14c412a3-bcdc-4a19-b769-310820125df0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.038039 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14c412a3-bcdc-4a19-b769-310820125df0-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.114932 4787 generic.go:334] "Generic (PLEG): container finished" podID="14c412a3-bcdc-4a19-b769-310820125df0" containerID="9fef4997b8b7315a153484f251256d79d41f4ad455d3e9db1fba2a85071cfd46" exitCode=0 Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.114986 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14c412a3-bcdc-4a19-b769-310820125df0","Type":"ContainerDied","Data":"9fef4997b8b7315a153484f251256d79d41f4ad455d3e9db1fba2a85071cfd46"} Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.115014 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14c412a3-bcdc-4a19-b769-310820125df0","Type":"ContainerDied","Data":"c5bfecd60b3009268f3c0a020b91d26523d48fc5748cc805fb4bad922ed5e8f5"} Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.115177 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.115239 4787 scope.go:117] "RemoveContainer" containerID="259adfed6bd06e27e49c4f41edf0c5edb0070ac71bdea39e66a2781d171c6aee" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.161326 4787 scope.go:117] "RemoveContainer" containerID="949b5aca8fd3ff42ce062ecb5c220a87429fed548318c3024c3a86168f30f764" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.161663 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.169441 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.191585 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:53:46 crc kubenswrapper[4787]: E1001 09:53:46.191970 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14c412a3-bcdc-4a19-b769-310820125df0" containerName="ceilometer-central-agent" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.191989 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="14c412a3-bcdc-4a19-b769-310820125df0" containerName="ceilometer-central-agent" Oct 01 09:53:46 crc kubenswrapper[4787]: E1001 09:53:46.192004 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14c412a3-bcdc-4a19-b769-310820125df0" containerName="sg-core" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.192011 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="14c412a3-bcdc-4a19-b769-310820125df0" containerName="sg-core" Oct 01 09:53:46 crc kubenswrapper[4787]: E1001 09:53:46.192026 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14c412a3-bcdc-4a19-b769-310820125df0" containerName="ceilometer-notification-agent" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.192036 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="14c412a3-bcdc-4a19-b769-310820125df0" containerName="ceilometer-notification-agent" Oct 01 09:53:46 crc kubenswrapper[4787]: E1001 09:53:46.192091 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14c412a3-bcdc-4a19-b769-310820125df0" containerName="proxy-httpd" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.192102 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="14c412a3-bcdc-4a19-b769-310820125df0" containerName="proxy-httpd" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.192276 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="14c412a3-bcdc-4a19-b769-310820125df0" containerName="sg-core" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.192302 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="14c412a3-bcdc-4a19-b769-310820125df0" containerName="proxy-httpd" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.192316 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="14c412a3-bcdc-4a19-b769-310820125df0" containerName="ceilometer-notification-agent" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.192324 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="14c412a3-bcdc-4a19-b769-310820125df0" containerName="ceilometer-central-agent" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.194023 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.198660 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.198938 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.202542 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.237344 4787 scope.go:117] "RemoveContainer" containerID="820b434e99203b0219ec476dc6a6a9fdfe43a8a8957692874fe905ef8dc37585" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.271374 4787 scope.go:117] "RemoveContainer" containerID="9fef4997b8b7315a153484f251256d79d41f4ad455d3e9db1fba2a85071cfd46" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.294726 4787 scope.go:117] "RemoveContainer" containerID="259adfed6bd06e27e49c4f41edf0c5edb0070ac71bdea39e66a2781d171c6aee" Oct 01 09:53:46 crc kubenswrapper[4787]: E1001 09:53:46.295232 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"259adfed6bd06e27e49c4f41edf0c5edb0070ac71bdea39e66a2781d171c6aee\": container with ID starting with 259adfed6bd06e27e49c4f41edf0c5edb0070ac71bdea39e66a2781d171c6aee not found: ID does not exist" containerID="259adfed6bd06e27e49c4f41edf0c5edb0070ac71bdea39e66a2781d171c6aee" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.295276 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"259adfed6bd06e27e49c4f41edf0c5edb0070ac71bdea39e66a2781d171c6aee"} err="failed to get container status \"259adfed6bd06e27e49c4f41edf0c5edb0070ac71bdea39e66a2781d171c6aee\": rpc error: code = NotFound desc = could not find container \"259adfed6bd06e27e49c4f41edf0c5edb0070ac71bdea39e66a2781d171c6aee\": container with ID starting with 259adfed6bd06e27e49c4f41edf0c5edb0070ac71bdea39e66a2781d171c6aee not found: ID does not exist" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.295304 4787 scope.go:117] "RemoveContainer" containerID="949b5aca8fd3ff42ce062ecb5c220a87429fed548318c3024c3a86168f30f764" Oct 01 09:53:46 crc kubenswrapper[4787]: E1001 09:53:46.295528 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"949b5aca8fd3ff42ce062ecb5c220a87429fed548318c3024c3a86168f30f764\": container with ID starting with 949b5aca8fd3ff42ce062ecb5c220a87429fed548318c3024c3a86168f30f764 not found: ID does not exist" containerID="949b5aca8fd3ff42ce062ecb5c220a87429fed548318c3024c3a86168f30f764" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.295553 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"949b5aca8fd3ff42ce062ecb5c220a87429fed548318c3024c3a86168f30f764"} err="failed to get container status \"949b5aca8fd3ff42ce062ecb5c220a87429fed548318c3024c3a86168f30f764\": rpc error: code = NotFound desc = could not find container \"949b5aca8fd3ff42ce062ecb5c220a87429fed548318c3024c3a86168f30f764\": container with ID starting with 949b5aca8fd3ff42ce062ecb5c220a87429fed548318c3024c3a86168f30f764 not found: ID does not exist" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.295568 4787 scope.go:117] "RemoveContainer" containerID="820b434e99203b0219ec476dc6a6a9fdfe43a8a8957692874fe905ef8dc37585" Oct 01 09:53:46 crc kubenswrapper[4787]: E1001 09:53:46.295900 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"820b434e99203b0219ec476dc6a6a9fdfe43a8a8957692874fe905ef8dc37585\": container with ID starting with 820b434e99203b0219ec476dc6a6a9fdfe43a8a8957692874fe905ef8dc37585 not found: ID does not exist" containerID="820b434e99203b0219ec476dc6a6a9fdfe43a8a8957692874fe905ef8dc37585" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.295924 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"820b434e99203b0219ec476dc6a6a9fdfe43a8a8957692874fe905ef8dc37585"} err="failed to get container status \"820b434e99203b0219ec476dc6a6a9fdfe43a8a8957692874fe905ef8dc37585\": rpc error: code = NotFound desc = could not find container \"820b434e99203b0219ec476dc6a6a9fdfe43a8a8957692874fe905ef8dc37585\": container with ID starting with 820b434e99203b0219ec476dc6a6a9fdfe43a8a8957692874fe905ef8dc37585 not found: ID does not exist" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.295938 4787 scope.go:117] "RemoveContainer" containerID="9fef4997b8b7315a153484f251256d79d41f4ad455d3e9db1fba2a85071cfd46" Oct 01 09:53:46 crc kubenswrapper[4787]: E1001 09:53:46.296361 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fef4997b8b7315a153484f251256d79d41f4ad455d3e9db1fba2a85071cfd46\": container with ID starting with 9fef4997b8b7315a153484f251256d79d41f4ad455d3e9db1fba2a85071cfd46 not found: ID does not exist" containerID="9fef4997b8b7315a153484f251256d79d41f4ad455d3e9db1fba2a85071cfd46" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.296384 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fef4997b8b7315a153484f251256d79d41f4ad455d3e9db1fba2a85071cfd46"} err="failed to get container status \"9fef4997b8b7315a153484f251256d79d41f4ad455d3e9db1fba2a85071cfd46\": rpc error: code = NotFound desc = could not find container \"9fef4997b8b7315a153484f251256d79d41f4ad455d3e9db1fba2a85071cfd46\": container with ID starting with 9fef4997b8b7315a153484f251256d79d41f4ad455d3e9db1fba2a85071cfd46 not found: ID does not exist" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.344479 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qd4ps\" (UniqueName: \"kubernetes.io/projected/c8ba0112-0496-47a6-892b-67de901a1dc7-kube-api-access-qd4ps\") pod \"ceilometer-0\" (UID: \"c8ba0112-0496-47a6-892b-67de901a1dc7\") " pod="openstack/ceilometer-0" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.344560 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8ba0112-0496-47a6-892b-67de901a1dc7-run-httpd\") pod \"ceilometer-0\" (UID: \"c8ba0112-0496-47a6-892b-67de901a1dc7\") " pod="openstack/ceilometer-0" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.344585 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c8ba0112-0496-47a6-892b-67de901a1dc7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c8ba0112-0496-47a6-892b-67de901a1dc7\") " pod="openstack/ceilometer-0" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.344797 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8ba0112-0496-47a6-892b-67de901a1dc7-log-httpd\") pod \"ceilometer-0\" (UID: \"c8ba0112-0496-47a6-892b-67de901a1dc7\") " pod="openstack/ceilometer-0" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.345237 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8ba0112-0496-47a6-892b-67de901a1dc7-config-data\") pod \"ceilometer-0\" (UID: \"c8ba0112-0496-47a6-892b-67de901a1dc7\") " pod="openstack/ceilometer-0" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.345278 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8ba0112-0496-47a6-892b-67de901a1dc7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c8ba0112-0496-47a6-892b-67de901a1dc7\") " pod="openstack/ceilometer-0" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.345570 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8ba0112-0496-47a6-892b-67de901a1dc7-scripts\") pod \"ceilometer-0\" (UID: \"c8ba0112-0496-47a6-892b-67de901a1dc7\") " pod="openstack/ceilometer-0" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.447525 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8ba0112-0496-47a6-892b-67de901a1dc7-log-httpd\") pod \"ceilometer-0\" (UID: \"c8ba0112-0496-47a6-892b-67de901a1dc7\") " pod="openstack/ceilometer-0" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.447586 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8ba0112-0496-47a6-892b-67de901a1dc7-config-data\") pod \"ceilometer-0\" (UID: \"c8ba0112-0496-47a6-892b-67de901a1dc7\") " pod="openstack/ceilometer-0" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.447612 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8ba0112-0496-47a6-892b-67de901a1dc7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c8ba0112-0496-47a6-892b-67de901a1dc7\") " pod="openstack/ceilometer-0" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.447700 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8ba0112-0496-47a6-892b-67de901a1dc7-scripts\") pod \"ceilometer-0\" (UID: \"c8ba0112-0496-47a6-892b-67de901a1dc7\") " pod="openstack/ceilometer-0" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.447750 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qd4ps\" (UniqueName: \"kubernetes.io/projected/c8ba0112-0496-47a6-892b-67de901a1dc7-kube-api-access-qd4ps\") pod \"ceilometer-0\" (UID: \"c8ba0112-0496-47a6-892b-67de901a1dc7\") " pod="openstack/ceilometer-0" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.447795 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8ba0112-0496-47a6-892b-67de901a1dc7-run-httpd\") pod \"ceilometer-0\" (UID: \"c8ba0112-0496-47a6-892b-67de901a1dc7\") " pod="openstack/ceilometer-0" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.447811 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c8ba0112-0496-47a6-892b-67de901a1dc7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c8ba0112-0496-47a6-892b-67de901a1dc7\") " pod="openstack/ceilometer-0" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.448706 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8ba0112-0496-47a6-892b-67de901a1dc7-log-httpd\") pod \"ceilometer-0\" (UID: \"c8ba0112-0496-47a6-892b-67de901a1dc7\") " pod="openstack/ceilometer-0" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.450134 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8ba0112-0496-47a6-892b-67de901a1dc7-run-httpd\") pod \"ceilometer-0\" (UID: \"c8ba0112-0496-47a6-892b-67de901a1dc7\") " pod="openstack/ceilometer-0" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.453934 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8ba0112-0496-47a6-892b-67de901a1dc7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c8ba0112-0496-47a6-892b-67de901a1dc7\") " pod="openstack/ceilometer-0" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.455620 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8ba0112-0496-47a6-892b-67de901a1dc7-config-data\") pod \"ceilometer-0\" (UID: \"c8ba0112-0496-47a6-892b-67de901a1dc7\") " pod="openstack/ceilometer-0" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.458313 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8ba0112-0496-47a6-892b-67de901a1dc7-scripts\") pod \"ceilometer-0\" (UID: \"c8ba0112-0496-47a6-892b-67de901a1dc7\") " pod="openstack/ceilometer-0" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.465527 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c8ba0112-0496-47a6-892b-67de901a1dc7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c8ba0112-0496-47a6-892b-67de901a1dc7\") " pod="openstack/ceilometer-0" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.469250 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qd4ps\" (UniqueName: \"kubernetes.io/projected/c8ba0112-0496-47a6-892b-67de901a1dc7-kube-api-access-qd4ps\") pod \"ceilometer-0\" (UID: \"c8ba0112-0496-47a6-892b-67de901a1dc7\") " pod="openstack/ceilometer-0" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.516172 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.537462 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14c412a3-bcdc-4a19-b769-310820125df0" path="/var/lib/kubelet/pods/14c412a3-bcdc-4a19-b769-310820125df0/volumes" Oct 01 09:53:46 crc kubenswrapper[4787]: I1001 09:53:46.946475 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:53:46 crc kubenswrapper[4787]: W1001 09:53:46.947214 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc8ba0112_0496_47a6_892b_67de901a1dc7.slice/crio-325ac7878eebfb5539331b3a4af21a740102a559f5d620c116e601c1c6602153 WatchSource:0}: Error finding container 325ac7878eebfb5539331b3a4af21a740102a559f5d620c116e601c1c6602153: Status 404 returned error can't find the container with id 325ac7878eebfb5539331b3a4af21a740102a559f5d620c116e601c1c6602153 Oct 01 09:53:47 crc kubenswrapper[4787]: I1001 09:53:47.126994 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8ba0112-0496-47a6-892b-67de901a1dc7","Type":"ContainerStarted","Data":"325ac7878eebfb5539331b3a4af21a740102a559f5d620c116e601c1c6602153"} Oct 01 09:53:47 crc kubenswrapper[4787]: I1001 09:53:47.557179 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 01 09:53:54 crc kubenswrapper[4787]: I1001 09:53:54.199568 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8ba0112-0496-47a6-892b-67de901a1dc7","Type":"ContainerStarted","Data":"f32a6a29f32059786df18ef5599f60b5ddef12166f239c441a4fac09b5fadff8"} Oct 01 09:53:54 crc kubenswrapper[4787]: I1001 09:53:54.200228 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8ba0112-0496-47a6-892b-67de901a1dc7","Type":"ContainerStarted","Data":"ebd86d6ec23a4190ed72829112316979d7dca0056c7910569f979ecb8bebbeba"} Oct 01 09:53:54 crc kubenswrapper[4787]: I1001 09:53:54.201451 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-tj64b" event={"ID":"9821f091-0513-4370-beb7-8239eab74972","Type":"ContainerStarted","Data":"669a8eed122a993a2d627ab48625d872ec440697cf9206734d572865c06fccbf"} Oct 01 09:53:54 crc kubenswrapper[4787]: I1001 09:53:54.226057 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-tj64b" podStartSLOduration=2.429128206 podStartE2EDuration="11.226038039s" podCreationTimestamp="2025-10-01 09:53:43 +0000 UTC" firstStartedPulling="2025-10-01 09:53:44.188678326 +0000 UTC m=+1056.303822483" lastFinishedPulling="2025-10-01 09:53:52.985588159 +0000 UTC m=+1065.100732316" observedRunningTime="2025-10-01 09:53:54.220431717 +0000 UTC m=+1066.335575884" watchObservedRunningTime="2025-10-01 09:53:54.226038039 +0000 UTC m=+1066.341182196" Oct 01 09:53:55 crc kubenswrapper[4787]: I1001 09:53:55.222694 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8ba0112-0496-47a6-892b-67de901a1dc7","Type":"ContainerStarted","Data":"92672e09e5a61f7d864f2f68a2142fb28ce00f9224ba599a459c8287c4b3a958"} Oct 01 09:53:56 crc kubenswrapper[4787]: I1001 09:53:56.233886 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8ba0112-0496-47a6-892b-67de901a1dc7","Type":"ContainerStarted","Data":"e13f12c5c3678f856ea6fc07b93fa3285dc5454ea50d6b707d768f3283cdd871"} Oct 01 09:53:56 crc kubenswrapper[4787]: I1001 09:53:56.234537 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 01 09:53:56 crc kubenswrapper[4787]: I1001 09:53:56.257875 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.242237454 podStartE2EDuration="10.257829015s" podCreationTimestamp="2025-10-01 09:53:46 +0000 UTC" firstStartedPulling="2025-10-01 09:53:46.951551374 +0000 UTC m=+1059.066695531" lastFinishedPulling="2025-10-01 09:53:55.967142895 +0000 UTC m=+1068.082287092" observedRunningTime="2025-10-01 09:53:56.25723522 +0000 UTC m=+1068.372379377" watchObservedRunningTime="2025-10-01 09:53:56.257829015 +0000 UTC m=+1068.372973172" Oct 01 09:53:58 crc kubenswrapper[4787]: I1001 09:53:58.599343 4787 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod7983e213-c317-42b2-8944-85bdb78dac8f"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod7983e213-c317-42b2-8944-85bdb78dac8f] : Timed out while waiting for systemd to remove kubepods-besteffort-pod7983e213_c317_42b2_8944_85bdb78dac8f.slice" Oct 01 09:54:01 crc kubenswrapper[4787]: I1001 09:54:01.528777 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:54:01 crc kubenswrapper[4787]: I1001 09:54:01.529458 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c8ba0112-0496-47a6-892b-67de901a1dc7" containerName="proxy-httpd" containerID="cri-o://e13f12c5c3678f856ea6fc07b93fa3285dc5454ea50d6b707d768f3283cdd871" gracePeriod=30 Oct 01 09:54:01 crc kubenswrapper[4787]: I1001 09:54:01.529459 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c8ba0112-0496-47a6-892b-67de901a1dc7" containerName="sg-core" containerID="cri-o://92672e09e5a61f7d864f2f68a2142fb28ce00f9224ba599a459c8287c4b3a958" gracePeriod=30 Oct 01 09:54:01 crc kubenswrapper[4787]: I1001 09:54:01.529537 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c8ba0112-0496-47a6-892b-67de901a1dc7" containerName="ceilometer-notification-agent" containerID="cri-o://f32a6a29f32059786df18ef5599f60b5ddef12166f239c441a4fac09b5fadff8" gracePeriod=30 Oct 01 09:54:01 crc kubenswrapper[4787]: I1001 09:54:01.529423 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c8ba0112-0496-47a6-892b-67de901a1dc7" containerName="ceilometer-central-agent" containerID="cri-o://ebd86d6ec23a4190ed72829112316979d7dca0056c7910569f979ecb8bebbeba" gracePeriod=30 Oct 01 09:54:02 crc kubenswrapper[4787]: I1001 09:54:02.296930 4787 generic.go:334] "Generic (PLEG): container finished" podID="c8ba0112-0496-47a6-892b-67de901a1dc7" containerID="e13f12c5c3678f856ea6fc07b93fa3285dc5454ea50d6b707d768f3283cdd871" exitCode=0 Oct 01 09:54:02 crc kubenswrapper[4787]: I1001 09:54:02.297289 4787 generic.go:334] "Generic (PLEG): container finished" podID="c8ba0112-0496-47a6-892b-67de901a1dc7" containerID="92672e09e5a61f7d864f2f68a2142fb28ce00f9224ba599a459c8287c4b3a958" exitCode=2 Oct 01 09:54:02 crc kubenswrapper[4787]: I1001 09:54:02.297297 4787 generic.go:334] "Generic (PLEG): container finished" podID="c8ba0112-0496-47a6-892b-67de901a1dc7" containerID="f32a6a29f32059786df18ef5599f60b5ddef12166f239c441a4fac09b5fadff8" exitCode=0 Oct 01 09:54:02 crc kubenswrapper[4787]: I1001 09:54:02.297304 4787 generic.go:334] "Generic (PLEG): container finished" podID="c8ba0112-0496-47a6-892b-67de901a1dc7" containerID="ebd86d6ec23a4190ed72829112316979d7dca0056c7910569f979ecb8bebbeba" exitCode=0 Oct 01 09:54:02 crc kubenswrapper[4787]: I1001 09:54:02.297324 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8ba0112-0496-47a6-892b-67de901a1dc7","Type":"ContainerDied","Data":"e13f12c5c3678f856ea6fc07b93fa3285dc5454ea50d6b707d768f3283cdd871"} Oct 01 09:54:02 crc kubenswrapper[4787]: I1001 09:54:02.297350 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8ba0112-0496-47a6-892b-67de901a1dc7","Type":"ContainerDied","Data":"92672e09e5a61f7d864f2f68a2142fb28ce00f9224ba599a459c8287c4b3a958"} Oct 01 09:54:02 crc kubenswrapper[4787]: I1001 09:54:02.297359 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8ba0112-0496-47a6-892b-67de901a1dc7","Type":"ContainerDied","Data":"f32a6a29f32059786df18ef5599f60b5ddef12166f239c441a4fac09b5fadff8"} Oct 01 09:54:02 crc kubenswrapper[4787]: I1001 09:54:02.297368 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8ba0112-0496-47a6-892b-67de901a1dc7","Type":"ContainerDied","Data":"ebd86d6ec23a4190ed72829112316979d7dca0056c7910569f979ecb8bebbeba"} Oct 01 09:54:02 crc kubenswrapper[4787]: I1001 09:54:02.425795 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:54:02 crc kubenswrapper[4787]: I1001 09:54:02.584727 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8ba0112-0496-47a6-892b-67de901a1dc7-combined-ca-bundle\") pod \"c8ba0112-0496-47a6-892b-67de901a1dc7\" (UID: \"c8ba0112-0496-47a6-892b-67de901a1dc7\") " Oct 01 09:54:02 crc kubenswrapper[4787]: I1001 09:54:02.584792 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8ba0112-0496-47a6-892b-67de901a1dc7-config-data\") pod \"c8ba0112-0496-47a6-892b-67de901a1dc7\" (UID: \"c8ba0112-0496-47a6-892b-67de901a1dc7\") " Oct 01 09:54:02 crc kubenswrapper[4787]: I1001 09:54:02.584876 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c8ba0112-0496-47a6-892b-67de901a1dc7-sg-core-conf-yaml\") pod \"c8ba0112-0496-47a6-892b-67de901a1dc7\" (UID: \"c8ba0112-0496-47a6-892b-67de901a1dc7\") " Oct 01 09:54:02 crc kubenswrapper[4787]: I1001 09:54:02.584932 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qd4ps\" (UniqueName: \"kubernetes.io/projected/c8ba0112-0496-47a6-892b-67de901a1dc7-kube-api-access-qd4ps\") pod \"c8ba0112-0496-47a6-892b-67de901a1dc7\" (UID: \"c8ba0112-0496-47a6-892b-67de901a1dc7\") " Oct 01 09:54:02 crc kubenswrapper[4787]: I1001 09:54:02.584977 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8ba0112-0496-47a6-892b-67de901a1dc7-run-httpd\") pod \"c8ba0112-0496-47a6-892b-67de901a1dc7\" (UID: \"c8ba0112-0496-47a6-892b-67de901a1dc7\") " Oct 01 09:54:02 crc kubenswrapper[4787]: I1001 09:54:02.584998 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8ba0112-0496-47a6-892b-67de901a1dc7-scripts\") pod \"c8ba0112-0496-47a6-892b-67de901a1dc7\" (UID: \"c8ba0112-0496-47a6-892b-67de901a1dc7\") " Oct 01 09:54:02 crc kubenswrapper[4787]: I1001 09:54:02.585032 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8ba0112-0496-47a6-892b-67de901a1dc7-log-httpd\") pod \"c8ba0112-0496-47a6-892b-67de901a1dc7\" (UID: \"c8ba0112-0496-47a6-892b-67de901a1dc7\") " Oct 01 09:54:02 crc kubenswrapper[4787]: I1001 09:54:02.586122 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8ba0112-0496-47a6-892b-67de901a1dc7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c8ba0112-0496-47a6-892b-67de901a1dc7" (UID: "c8ba0112-0496-47a6-892b-67de901a1dc7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:54:02 crc kubenswrapper[4787]: I1001 09:54:02.586629 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8ba0112-0496-47a6-892b-67de901a1dc7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c8ba0112-0496-47a6-892b-67de901a1dc7" (UID: "c8ba0112-0496-47a6-892b-67de901a1dc7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:54:02 crc kubenswrapper[4787]: I1001 09:54:02.593238 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8ba0112-0496-47a6-892b-67de901a1dc7-kube-api-access-qd4ps" (OuterVolumeSpecName: "kube-api-access-qd4ps") pod "c8ba0112-0496-47a6-892b-67de901a1dc7" (UID: "c8ba0112-0496-47a6-892b-67de901a1dc7"). InnerVolumeSpecName "kube-api-access-qd4ps". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:54:02 crc kubenswrapper[4787]: I1001 09:54:02.593236 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8ba0112-0496-47a6-892b-67de901a1dc7-scripts" (OuterVolumeSpecName: "scripts") pod "c8ba0112-0496-47a6-892b-67de901a1dc7" (UID: "c8ba0112-0496-47a6-892b-67de901a1dc7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:54:02 crc kubenswrapper[4787]: I1001 09:54:02.624891 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8ba0112-0496-47a6-892b-67de901a1dc7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c8ba0112-0496-47a6-892b-67de901a1dc7" (UID: "c8ba0112-0496-47a6-892b-67de901a1dc7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:54:02 crc kubenswrapper[4787]: I1001 09:54:02.687893 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8ba0112-0496-47a6-892b-67de901a1dc7-config-data" (OuterVolumeSpecName: "config-data") pod "c8ba0112-0496-47a6-892b-67de901a1dc7" (UID: "c8ba0112-0496-47a6-892b-67de901a1dc7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:54:02 crc kubenswrapper[4787]: I1001 09:54:02.688556 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8ba0112-0496-47a6-892b-67de901a1dc7-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:02 crc kubenswrapper[4787]: I1001 09:54:02.688601 4787 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8ba0112-0496-47a6-892b-67de901a1dc7-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:02 crc kubenswrapper[4787]: I1001 09:54:02.688614 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8ba0112-0496-47a6-892b-67de901a1dc7-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:02 crc kubenswrapper[4787]: I1001 09:54:02.688626 4787 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c8ba0112-0496-47a6-892b-67de901a1dc7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:02 crc kubenswrapper[4787]: I1001 09:54:02.688641 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qd4ps\" (UniqueName: \"kubernetes.io/projected/c8ba0112-0496-47a6-892b-67de901a1dc7-kube-api-access-qd4ps\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:02 crc kubenswrapper[4787]: I1001 09:54:02.688654 4787 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8ba0112-0496-47a6-892b-67de901a1dc7-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:02 crc kubenswrapper[4787]: I1001 09:54:02.692604 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8ba0112-0496-47a6-892b-67de901a1dc7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c8ba0112-0496-47a6-892b-67de901a1dc7" (UID: "c8ba0112-0496-47a6-892b-67de901a1dc7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:54:02 crc kubenswrapper[4787]: I1001 09:54:02.789891 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8ba0112-0496-47a6-892b-67de901a1dc7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.315007 4787 generic.go:334] "Generic (PLEG): container finished" podID="9821f091-0513-4370-beb7-8239eab74972" containerID="669a8eed122a993a2d627ab48625d872ec440697cf9206734d572865c06fccbf" exitCode=0 Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.315149 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-tj64b" event={"ID":"9821f091-0513-4370-beb7-8239eab74972","Type":"ContainerDied","Data":"669a8eed122a993a2d627ab48625d872ec440697cf9206734d572865c06fccbf"} Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.318553 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8ba0112-0496-47a6-892b-67de901a1dc7","Type":"ContainerDied","Data":"325ac7878eebfb5539331b3a4af21a740102a559f5d620c116e601c1c6602153"} Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.318622 4787 scope.go:117] "RemoveContainer" containerID="e13f12c5c3678f856ea6fc07b93fa3285dc5454ea50d6b707d768f3283cdd871" Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.318659 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.352749 4787 scope.go:117] "RemoveContainer" containerID="92672e09e5a61f7d864f2f68a2142fb28ce00f9224ba599a459c8287c4b3a958" Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.370733 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.381642 4787 scope.go:117] "RemoveContainer" containerID="f32a6a29f32059786df18ef5599f60b5ddef12166f239c441a4fac09b5fadff8" Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.397067 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.420342 4787 scope.go:117] "RemoveContainer" containerID="ebd86d6ec23a4190ed72829112316979d7dca0056c7910569f979ecb8bebbeba" Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.421021 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:54:03 crc kubenswrapper[4787]: E1001 09:54:03.421725 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8ba0112-0496-47a6-892b-67de901a1dc7" containerName="proxy-httpd" Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.421757 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8ba0112-0496-47a6-892b-67de901a1dc7" containerName="proxy-httpd" Oct 01 09:54:03 crc kubenswrapper[4787]: E1001 09:54:03.421782 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8ba0112-0496-47a6-892b-67de901a1dc7" containerName="ceilometer-central-agent" Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.421794 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8ba0112-0496-47a6-892b-67de901a1dc7" containerName="ceilometer-central-agent" Oct 01 09:54:03 crc kubenswrapper[4787]: E1001 09:54:03.421848 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8ba0112-0496-47a6-892b-67de901a1dc7" containerName="sg-core" Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.421860 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8ba0112-0496-47a6-892b-67de901a1dc7" containerName="sg-core" Oct 01 09:54:03 crc kubenswrapper[4787]: E1001 09:54:03.421899 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8ba0112-0496-47a6-892b-67de901a1dc7" containerName="ceilometer-notification-agent" Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.421914 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8ba0112-0496-47a6-892b-67de901a1dc7" containerName="ceilometer-notification-agent" Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.422273 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8ba0112-0496-47a6-892b-67de901a1dc7" containerName="proxy-httpd" Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.422327 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8ba0112-0496-47a6-892b-67de901a1dc7" containerName="ceilometer-notification-agent" Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.422353 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8ba0112-0496-47a6-892b-67de901a1dc7" containerName="ceilometer-central-agent" Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.422376 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8ba0112-0496-47a6-892b-67de901a1dc7" containerName="sg-core" Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.425443 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.428790 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.428823 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.431800 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.509323 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-scripts\") pod \"ceilometer-0\" (UID: \"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5\") " pod="openstack/ceilometer-0" Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.509400 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5\") " pod="openstack/ceilometer-0" Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.509601 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grwtt\" (UniqueName: \"kubernetes.io/projected/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-kube-api-access-grwtt\") pod \"ceilometer-0\" (UID: \"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5\") " pod="openstack/ceilometer-0" Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.509732 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-log-httpd\") pod \"ceilometer-0\" (UID: \"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5\") " pod="openstack/ceilometer-0" Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.509889 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-config-data\") pod \"ceilometer-0\" (UID: \"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5\") " pod="openstack/ceilometer-0" Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.509980 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5\") " pod="openstack/ceilometer-0" Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.510197 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-run-httpd\") pod \"ceilometer-0\" (UID: \"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5\") " pod="openstack/ceilometer-0" Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.611867 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grwtt\" (UniqueName: \"kubernetes.io/projected/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-kube-api-access-grwtt\") pod \"ceilometer-0\" (UID: \"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5\") " pod="openstack/ceilometer-0" Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.611954 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-log-httpd\") pod \"ceilometer-0\" (UID: \"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5\") " pod="openstack/ceilometer-0" Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.612013 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-config-data\") pod \"ceilometer-0\" (UID: \"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5\") " pod="openstack/ceilometer-0" Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.612045 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5\") " pod="openstack/ceilometer-0" Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.612134 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-run-httpd\") pod \"ceilometer-0\" (UID: \"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5\") " pod="openstack/ceilometer-0" Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.612256 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-scripts\") pod \"ceilometer-0\" (UID: \"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5\") " pod="openstack/ceilometer-0" Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.612299 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5\") " pod="openstack/ceilometer-0" Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.613555 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-log-httpd\") pod \"ceilometer-0\" (UID: \"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5\") " pod="openstack/ceilometer-0" Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.614542 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-run-httpd\") pod \"ceilometer-0\" (UID: \"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5\") " pod="openstack/ceilometer-0" Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.616541 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5\") " pod="openstack/ceilometer-0" Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.617864 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5\") " pod="openstack/ceilometer-0" Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.618607 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-config-data\") pod \"ceilometer-0\" (UID: \"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5\") " pod="openstack/ceilometer-0" Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.634033 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-scripts\") pod \"ceilometer-0\" (UID: \"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5\") " pod="openstack/ceilometer-0" Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.645816 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grwtt\" (UniqueName: \"kubernetes.io/projected/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-kube-api-access-grwtt\") pod \"ceilometer-0\" (UID: \"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5\") " pod="openstack/ceilometer-0" Oct 01 09:54:03 crc kubenswrapper[4787]: I1001 09:54:03.757395 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:54:04 crc kubenswrapper[4787]: I1001 09:54:04.213489 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:54:04 crc kubenswrapper[4787]: I1001 09:54:04.329988 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5","Type":"ContainerStarted","Data":"b3e27dd3d23025dfe522f4d751d860646c014af5bc392ca7241cd4bdf30c3149"} Oct 01 09:54:04 crc kubenswrapper[4787]: I1001 09:54:04.536908 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8ba0112-0496-47a6-892b-67de901a1dc7" path="/var/lib/kubelet/pods/c8ba0112-0496-47a6-892b-67de901a1dc7/volumes" Oct 01 09:54:04 crc kubenswrapper[4787]: I1001 09:54:04.672488 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-tj64b" Oct 01 09:54:04 crc kubenswrapper[4787]: I1001 09:54:04.737230 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9821f091-0513-4370-beb7-8239eab74972-combined-ca-bundle\") pod \"9821f091-0513-4370-beb7-8239eab74972\" (UID: \"9821f091-0513-4370-beb7-8239eab74972\") " Oct 01 09:54:04 crc kubenswrapper[4787]: I1001 09:54:04.737786 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9821f091-0513-4370-beb7-8239eab74972-config-data\") pod \"9821f091-0513-4370-beb7-8239eab74972\" (UID: \"9821f091-0513-4370-beb7-8239eab74972\") " Oct 01 09:54:04 crc kubenswrapper[4787]: I1001 09:54:04.737939 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9821f091-0513-4370-beb7-8239eab74972-scripts\") pod \"9821f091-0513-4370-beb7-8239eab74972\" (UID: \"9821f091-0513-4370-beb7-8239eab74972\") " Oct 01 09:54:04 crc kubenswrapper[4787]: I1001 09:54:04.737973 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lhbl5\" (UniqueName: \"kubernetes.io/projected/9821f091-0513-4370-beb7-8239eab74972-kube-api-access-lhbl5\") pod \"9821f091-0513-4370-beb7-8239eab74972\" (UID: \"9821f091-0513-4370-beb7-8239eab74972\") " Oct 01 09:54:04 crc kubenswrapper[4787]: I1001 09:54:04.743802 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9821f091-0513-4370-beb7-8239eab74972-kube-api-access-lhbl5" (OuterVolumeSpecName: "kube-api-access-lhbl5") pod "9821f091-0513-4370-beb7-8239eab74972" (UID: "9821f091-0513-4370-beb7-8239eab74972"). InnerVolumeSpecName "kube-api-access-lhbl5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:54:04 crc kubenswrapper[4787]: I1001 09:54:04.744452 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9821f091-0513-4370-beb7-8239eab74972-scripts" (OuterVolumeSpecName: "scripts") pod "9821f091-0513-4370-beb7-8239eab74972" (UID: "9821f091-0513-4370-beb7-8239eab74972"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:54:04 crc kubenswrapper[4787]: I1001 09:54:04.770953 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9821f091-0513-4370-beb7-8239eab74972-config-data" (OuterVolumeSpecName: "config-data") pod "9821f091-0513-4370-beb7-8239eab74972" (UID: "9821f091-0513-4370-beb7-8239eab74972"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:54:04 crc kubenswrapper[4787]: I1001 09:54:04.772882 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9821f091-0513-4370-beb7-8239eab74972-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9821f091-0513-4370-beb7-8239eab74972" (UID: "9821f091-0513-4370-beb7-8239eab74972"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:54:04 crc kubenswrapper[4787]: I1001 09:54:04.840347 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9821f091-0513-4370-beb7-8239eab74972-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:04 crc kubenswrapper[4787]: I1001 09:54:04.840392 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhbl5\" (UniqueName: \"kubernetes.io/projected/9821f091-0513-4370-beb7-8239eab74972-kube-api-access-lhbl5\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:04 crc kubenswrapper[4787]: I1001 09:54:04.840407 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9821f091-0513-4370-beb7-8239eab74972-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:04 crc kubenswrapper[4787]: I1001 09:54:04.840436 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9821f091-0513-4370-beb7-8239eab74972-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:05 crc kubenswrapper[4787]: I1001 09:54:05.341152 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-tj64b" event={"ID":"9821f091-0513-4370-beb7-8239eab74972","Type":"ContainerDied","Data":"cabafafa3aa175c401aef7fdcc6f183dfaba812c9865c6c1ba582206ef9f5688"} Oct 01 09:54:05 crc kubenswrapper[4787]: I1001 09:54:05.341204 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-tj64b" Oct 01 09:54:05 crc kubenswrapper[4787]: I1001 09:54:05.341210 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cabafafa3aa175c401aef7fdcc6f183dfaba812c9865c6c1ba582206ef9f5688" Oct 01 09:54:05 crc kubenswrapper[4787]: I1001 09:54:05.342839 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5","Type":"ContainerStarted","Data":"e338c2385f03560491eccb8129f3be389c57a3b6d485da5c158a9038e5530863"} Oct 01 09:54:05 crc kubenswrapper[4787]: I1001 09:54:05.479044 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 01 09:54:05 crc kubenswrapper[4787]: E1001 09:54:05.479586 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9821f091-0513-4370-beb7-8239eab74972" containerName="nova-cell0-conductor-db-sync" Oct 01 09:54:05 crc kubenswrapper[4787]: I1001 09:54:05.479609 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="9821f091-0513-4370-beb7-8239eab74972" containerName="nova-cell0-conductor-db-sync" Oct 01 09:54:05 crc kubenswrapper[4787]: I1001 09:54:05.479872 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="9821f091-0513-4370-beb7-8239eab74972" containerName="nova-cell0-conductor-db-sync" Oct 01 09:54:05 crc kubenswrapper[4787]: I1001 09:54:05.480747 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 01 09:54:05 crc kubenswrapper[4787]: I1001 09:54:05.486050 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 01 09:54:05 crc kubenswrapper[4787]: I1001 09:54:05.486264 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-hjcxb" Oct 01 09:54:05 crc kubenswrapper[4787]: I1001 09:54:05.488684 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 01 09:54:05 crc kubenswrapper[4787]: I1001 09:54:05.556462 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39f3448f-eee4-419b-a33b-8776fc5df599-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"39f3448f-eee4-419b-a33b-8776fc5df599\") " pod="openstack/nova-cell0-conductor-0" Oct 01 09:54:05 crc kubenswrapper[4787]: I1001 09:54:05.556640 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pndnf\" (UniqueName: \"kubernetes.io/projected/39f3448f-eee4-419b-a33b-8776fc5df599-kube-api-access-pndnf\") pod \"nova-cell0-conductor-0\" (UID: \"39f3448f-eee4-419b-a33b-8776fc5df599\") " pod="openstack/nova-cell0-conductor-0" Oct 01 09:54:05 crc kubenswrapper[4787]: I1001 09:54:05.556735 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39f3448f-eee4-419b-a33b-8776fc5df599-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"39f3448f-eee4-419b-a33b-8776fc5df599\") " pod="openstack/nova-cell0-conductor-0" Oct 01 09:54:05 crc kubenswrapper[4787]: E1001 09:54:05.618561 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9821f091_0513_4370_beb7_8239eab74972.slice/crio-cabafafa3aa175c401aef7fdcc6f183dfaba812c9865c6c1ba582206ef9f5688\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9821f091_0513_4370_beb7_8239eab74972.slice\": RecentStats: unable to find data in memory cache]" Oct 01 09:54:05 crc kubenswrapper[4787]: I1001 09:54:05.659770 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39f3448f-eee4-419b-a33b-8776fc5df599-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"39f3448f-eee4-419b-a33b-8776fc5df599\") " pod="openstack/nova-cell0-conductor-0" Oct 01 09:54:05 crc kubenswrapper[4787]: I1001 09:54:05.659880 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pndnf\" (UniqueName: \"kubernetes.io/projected/39f3448f-eee4-419b-a33b-8776fc5df599-kube-api-access-pndnf\") pod \"nova-cell0-conductor-0\" (UID: \"39f3448f-eee4-419b-a33b-8776fc5df599\") " pod="openstack/nova-cell0-conductor-0" Oct 01 09:54:05 crc kubenswrapper[4787]: I1001 09:54:05.659930 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39f3448f-eee4-419b-a33b-8776fc5df599-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"39f3448f-eee4-419b-a33b-8776fc5df599\") " pod="openstack/nova-cell0-conductor-0" Oct 01 09:54:05 crc kubenswrapper[4787]: I1001 09:54:05.664322 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39f3448f-eee4-419b-a33b-8776fc5df599-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"39f3448f-eee4-419b-a33b-8776fc5df599\") " pod="openstack/nova-cell0-conductor-0" Oct 01 09:54:05 crc kubenswrapper[4787]: I1001 09:54:05.664536 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39f3448f-eee4-419b-a33b-8776fc5df599-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"39f3448f-eee4-419b-a33b-8776fc5df599\") " pod="openstack/nova-cell0-conductor-0" Oct 01 09:54:05 crc kubenswrapper[4787]: I1001 09:54:05.688979 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pndnf\" (UniqueName: \"kubernetes.io/projected/39f3448f-eee4-419b-a33b-8776fc5df599-kube-api-access-pndnf\") pod \"nova-cell0-conductor-0\" (UID: \"39f3448f-eee4-419b-a33b-8776fc5df599\") " pod="openstack/nova-cell0-conductor-0" Oct 01 09:54:05 crc kubenswrapper[4787]: I1001 09:54:05.808800 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 01 09:54:06 crc kubenswrapper[4787]: I1001 09:54:06.333007 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 01 09:54:06 crc kubenswrapper[4787]: W1001 09:54:06.337228 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39f3448f_eee4_419b_a33b_8776fc5df599.slice/crio-871f7a8d62db27805f95eb83423816c5e463952aa6b010fa0762ddb3ac6a602c WatchSource:0}: Error finding container 871f7a8d62db27805f95eb83423816c5e463952aa6b010fa0762ddb3ac6a602c: Status 404 returned error can't find the container with id 871f7a8d62db27805f95eb83423816c5e463952aa6b010fa0762ddb3ac6a602c Oct 01 09:54:06 crc kubenswrapper[4787]: I1001 09:54:06.358646 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5","Type":"ContainerStarted","Data":"5ba6561fb8134d0a04c50d0e599b609eab84e3607f9c90d31673b339326eda2f"} Oct 01 09:54:06 crc kubenswrapper[4787]: I1001 09:54:06.359767 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"39f3448f-eee4-419b-a33b-8776fc5df599","Type":"ContainerStarted","Data":"871f7a8d62db27805f95eb83423816c5e463952aa6b010fa0762ddb3ac6a602c"} Oct 01 09:54:07 crc kubenswrapper[4787]: I1001 09:54:07.374551 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5","Type":"ContainerStarted","Data":"d1a6aa02188231959ecffe4fb8f1113d342324f03ac38e6084c61225e4e7579c"} Oct 01 09:54:07 crc kubenswrapper[4787]: I1001 09:54:07.376986 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"39f3448f-eee4-419b-a33b-8776fc5df599","Type":"ContainerStarted","Data":"5193bc095243d6d01b7013d9cc087d60070f2d18939b57a4e6262597da420f88"} Oct 01 09:54:07 crc kubenswrapper[4787]: I1001 09:54:07.400744 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.400724361 podStartE2EDuration="2.400724361s" podCreationTimestamp="2025-10-01 09:54:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:54:07.392459332 +0000 UTC m=+1079.507603489" watchObservedRunningTime="2025-10-01 09:54:07.400724361 +0000 UTC m=+1079.515868518" Oct 01 09:54:08 crc kubenswrapper[4787]: I1001 09:54:08.388161 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5","Type":"ContainerStarted","Data":"ada12ec4a7e7f6e31663ea7ee1e674967e3d7259c5c0ea8efc31cc2b343291e1"} Oct 01 09:54:08 crc kubenswrapper[4787]: I1001 09:54:08.389337 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 01 09:54:08 crc kubenswrapper[4787]: I1001 09:54:08.426400 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.027030037 podStartE2EDuration="5.426375781s" podCreationTimestamp="2025-10-01 09:54:03 +0000 UTC" firstStartedPulling="2025-10-01 09:54:04.220186387 +0000 UTC m=+1076.335330544" lastFinishedPulling="2025-10-01 09:54:07.619532131 +0000 UTC m=+1079.734676288" observedRunningTime="2025-10-01 09:54:08.42356422 +0000 UTC m=+1080.538708377" watchObservedRunningTime="2025-10-01 09:54:08.426375781 +0000 UTC m=+1080.541519938" Oct 01 09:54:09 crc kubenswrapper[4787]: I1001 09:54:09.398127 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 01 09:54:11 crc kubenswrapper[4787]: I1001 09:54:11.250206 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:54:11 crc kubenswrapper[4787]: I1001 09:54:11.250974 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:54:15 crc kubenswrapper[4787]: I1001 09:54:15.841331 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.334673 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-xglhv"] Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.336094 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-xglhv" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.342973 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-xglhv"] Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.343430 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.343642 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.382328 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0634a251-9095-454e-8a37-ff57c2ec40e1-config-data\") pod \"nova-cell0-cell-mapping-xglhv\" (UID: \"0634a251-9095-454e-8a37-ff57c2ec40e1\") " pod="openstack/nova-cell0-cell-mapping-xglhv" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.382977 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgllc\" (UniqueName: \"kubernetes.io/projected/0634a251-9095-454e-8a37-ff57c2ec40e1-kube-api-access-zgllc\") pod \"nova-cell0-cell-mapping-xglhv\" (UID: \"0634a251-9095-454e-8a37-ff57c2ec40e1\") " pod="openstack/nova-cell0-cell-mapping-xglhv" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.383067 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0634a251-9095-454e-8a37-ff57c2ec40e1-scripts\") pod \"nova-cell0-cell-mapping-xglhv\" (UID: \"0634a251-9095-454e-8a37-ff57c2ec40e1\") " pod="openstack/nova-cell0-cell-mapping-xglhv" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.383168 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0634a251-9095-454e-8a37-ff57c2ec40e1-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-xglhv\" (UID: \"0634a251-9095-454e-8a37-ff57c2ec40e1\") " pod="openstack/nova-cell0-cell-mapping-xglhv" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.487221 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgllc\" (UniqueName: \"kubernetes.io/projected/0634a251-9095-454e-8a37-ff57c2ec40e1-kube-api-access-zgllc\") pod \"nova-cell0-cell-mapping-xglhv\" (UID: \"0634a251-9095-454e-8a37-ff57c2ec40e1\") " pod="openstack/nova-cell0-cell-mapping-xglhv" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.487283 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0634a251-9095-454e-8a37-ff57c2ec40e1-scripts\") pod \"nova-cell0-cell-mapping-xglhv\" (UID: \"0634a251-9095-454e-8a37-ff57c2ec40e1\") " pod="openstack/nova-cell0-cell-mapping-xglhv" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.487314 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0634a251-9095-454e-8a37-ff57c2ec40e1-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-xglhv\" (UID: \"0634a251-9095-454e-8a37-ff57c2ec40e1\") " pod="openstack/nova-cell0-cell-mapping-xglhv" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.487361 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0634a251-9095-454e-8a37-ff57c2ec40e1-config-data\") pod \"nova-cell0-cell-mapping-xglhv\" (UID: \"0634a251-9095-454e-8a37-ff57c2ec40e1\") " pod="openstack/nova-cell0-cell-mapping-xglhv" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.496953 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0634a251-9095-454e-8a37-ff57c2ec40e1-config-data\") pod \"nova-cell0-cell-mapping-xglhv\" (UID: \"0634a251-9095-454e-8a37-ff57c2ec40e1\") " pod="openstack/nova-cell0-cell-mapping-xglhv" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.502131 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0634a251-9095-454e-8a37-ff57c2ec40e1-scripts\") pod \"nova-cell0-cell-mapping-xglhv\" (UID: \"0634a251-9095-454e-8a37-ff57c2ec40e1\") " pod="openstack/nova-cell0-cell-mapping-xglhv" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.502915 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0634a251-9095-454e-8a37-ff57c2ec40e1-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-xglhv\" (UID: \"0634a251-9095-454e-8a37-ff57c2ec40e1\") " pod="openstack/nova-cell0-cell-mapping-xglhv" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.531506 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgllc\" (UniqueName: \"kubernetes.io/projected/0634a251-9095-454e-8a37-ff57c2ec40e1-kube-api-access-zgllc\") pod \"nova-cell0-cell-mapping-xglhv\" (UID: \"0634a251-9095-454e-8a37-ff57c2ec40e1\") " pod="openstack/nova-cell0-cell-mapping-xglhv" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.541307 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.545996 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.547674 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.551329 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.589849 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54fpx\" (UniqueName: \"kubernetes.io/projected/13182f8e-43d0-40f7-b5a4-f11f34998d3d-kube-api-access-54fpx\") pod \"nova-cell1-novncproxy-0\" (UID: \"13182f8e-43d0-40f7-b5a4-f11f34998d3d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.589983 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13182f8e-43d0-40f7-b5a4-f11f34998d3d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"13182f8e-43d0-40f7-b5a4-f11f34998d3d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.590116 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13182f8e-43d0-40f7-b5a4-f11f34998d3d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"13182f8e-43d0-40f7-b5a4-f11f34998d3d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.622697 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.631694 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.644840 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.645128 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.674564 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-xglhv" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.700196 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13182f8e-43d0-40f7-b5a4-f11f34998d3d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"13182f8e-43d0-40f7-b5a4-f11f34998d3d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.700245 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06718a12-c4cb-43ed-be5f-0bee09272996-config-data\") pod \"nova-api-0\" (UID: \"06718a12-c4cb-43ed-be5f-0bee09272996\") " pod="openstack/nova-api-0" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.700274 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/06718a12-c4cb-43ed-be5f-0bee09272996-logs\") pod \"nova-api-0\" (UID: \"06718a12-c4cb-43ed-be5f-0bee09272996\") " pod="openstack/nova-api-0" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.700342 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhbgg\" (UniqueName: \"kubernetes.io/projected/06718a12-c4cb-43ed-be5f-0bee09272996-kube-api-access-zhbgg\") pod \"nova-api-0\" (UID: \"06718a12-c4cb-43ed-be5f-0bee09272996\") " pod="openstack/nova-api-0" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.700362 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13182f8e-43d0-40f7-b5a4-f11f34998d3d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"13182f8e-43d0-40f7-b5a4-f11f34998d3d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.700387 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06718a12-c4cb-43ed-be5f-0bee09272996-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"06718a12-c4cb-43ed-be5f-0bee09272996\") " pod="openstack/nova-api-0" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.700444 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54fpx\" (UniqueName: \"kubernetes.io/projected/13182f8e-43d0-40f7-b5a4-f11f34998d3d-kube-api-access-54fpx\") pod \"nova-cell1-novncproxy-0\" (UID: \"13182f8e-43d0-40f7-b5a4-f11f34998d3d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.708876 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13182f8e-43d0-40f7-b5a4-f11f34998d3d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"13182f8e-43d0-40f7-b5a4-f11f34998d3d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.725797 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13182f8e-43d0-40f7-b5a4-f11f34998d3d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"13182f8e-43d0-40f7-b5a4-f11f34998d3d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.753907 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54fpx\" (UniqueName: \"kubernetes.io/projected/13182f8e-43d0-40f7-b5a4-f11f34998d3d-kube-api-access-54fpx\") pod \"nova-cell1-novncproxy-0\" (UID: \"13182f8e-43d0-40f7-b5a4-f11f34998d3d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.811397 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhbgg\" (UniqueName: \"kubernetes.io/projected/06718a12-c4cb-43ed-be5f-0bee09272996-kube-api-access-zhbgg\") pod \"nova-api-0\" (UID: \"06718a12-c4cb-43ed-be5f-0bee09272996\") " pod="openstack/nova-api-0" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.811444 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06718a12-c4cb-43ed-be5f-0bee09272996-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"06718a12-c4cb-43ed-be5f-0bee09272996\") " pod="openstack/nova-api-0" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.811531 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06718a12-c4cb-43ed-be5f-0bee09272996-config-data\") pod \"nova-api-0\" (UID: \"06718a12-c4cb-43ed-be5f-0bee09272996\") " pod="openstack/nova-api-0" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.811551 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/06718a12-c4cb-43ed-be5f-0bee09272996-logs\") pod \"nova-api-0\" (UID: \"06718a12-c4cb-43ed-be5f-0bee09272996\") " pod="openstack/nova-api-0" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.811916 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/06718a12-c4cb-43ed-be5f-0bee09272996-logs\") pod \"nova-api-0\" (UID: \"06718a12-c4cb-43ed-be5f-0bee09272996\") " pod="openstack/nova-api-0" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.831872 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06718a12-c4cb-43ed-be5f-0bee09272996-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"06718a12-c4cb-43ed-be5f-0bee09272996\") " pod="openstack/nova-api-0" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.832501 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06718a12-c4cb-43ed-be5f-0bee09272996-config-data\") pod \"nova-api-0\" (UID: \"06718a12-c4cb-43ed-be5f-0bee09272996\") " pod="openstack/nova-api-0" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.886640 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.888653 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.895729 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhbgg\" (UniqueName: \"kubernetes.io/projected/06718a12-c4cb-43ed-be5f-0bee09272996-kube-api-access-zhbgg\") pod \"nova-api-0\" (UID: \"06718a12-c4cb-43ed-be5f-0bee09272996\") " pod="openstack/nova-api-0" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.906956 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.914409 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.918145 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c7de75f-5388-4bda-97ef-d3a17234001b-config-data\") pod \"nova-metadata-0\" (UID: \"2c7de75f-5388-4bda-97ef-d3a17234001b\") " pod="openstack/nova-metadata-0" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.918200 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c7de75f-5388-4bda-97ef-d3a17234001b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2c7de75f-5388-4bda-97ef-d3a17234001b\") " pod="openstack/nova-metadata-0" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.918249 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c7de75f-5388-4bda-97ef-d3a17234001b-logs\") pod \"nova-metadata-0\" (UID: \"2c7de75f-5388-4bda-97ef-d3a17234001b\") " pod="openstack/nova-metadata-0" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.918309 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccmqn\" (UniqueName: \"kubernetes.io/projected/2c7de75f-5388-4bda-97ef-d3a17234001b-kube-api-access-ccmqn\") pod \"nova-metadata-0\" (UID: \"2c7de75f-5388-4bda-97ef-d3a17234001b\") " pod="openstack/nova-metadata-0" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.919295 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.929889 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.949688 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.971011 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 09:54:16 crc kubenswrapper[4787]: I1001 09:54:16.986218 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.018413 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.020724 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c7de75f-5388-4bda-97ef-d3a17234001b-logs\") pod \"nova-metadata-0\" (UID: \"2c7de75f-5388-4bda-97ef-d3a17234001b\") " pod="openstack/nova-metadata-0" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.020811 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knchj\" (UniqueName: \"kubernetes.io/projected/e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4-kube-api-access-knchj\") pod \"nova-scheduler-0\" (UID: \"e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4\") " pod="openstack/nova-scheduler-0" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.020885 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccmqn\" (UniqueName: \"kubernetes.io/projected/2c7de75f-5388-4bda-97ef-d3a17234001b-kube-api-access-ccmqn\") pod \"nova-metadata-0\" (UID: \"2c7de75f-5388-4bda-97ef-d3a17234001b\") " pod="openstack/nova-metadata-0" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.020973 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4\") " pod="openstack/nova-scheduler-0" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.021034 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4-config-data\") pod \"nova-scheduler-0\" (UID: \"e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4\") " pod="openstack/nova-scheduler-0" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.021105 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c7de75f-5388-4bda-97ef-d3a17234001b-config-data\") pod \"nova-metadata-0\" (UID: \"2c7de75f-5388-4bda-97ef-d3a17234001b\") " pod="openstack/nova-metadata-0" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.021220 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c7de75f-5388-4bda-97ef-d3a17234001b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2c7de75f-5388-4bda-97ef-d3a17234001b\") " pod="openstack/nova-metadata-0" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.021712 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c7de75f-5388-4bda-97ef-d3a17234001b-logs\") pod \"nova-metadata-0\" (UID: \"2c7de75f-5388-4bda-97ef-d3a17234001b\") " pod="openstack/nova-metadata-0" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.034915 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c7de75f-5388-4bda-97ef-d3a17234001b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2c7de75f-5388-4bda-97ef-d3a17234001b\") " pod="openstack/nova-metadata-0" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.036257 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c7de75f-5388-4bda-97ef-d3a17234001b-config-data\") pod \"nova-metadata-0\" (UID: \"2c7de75f-5388-4bda-97ef-d3a17234001b\") " pod="openstack/nova-metadata-0" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.066962 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccmqn\" (UniqueName: \"kubernetes.io/projected/2c7de75f-5388-4bda-97ef-d3a17234001b-kube-api-access-ccmqn\") pod \"nova-metadata-0\" (UID: \"2c7de75f-5388-4bda-97ef-d3a17234001b\") " pod="openstack/nova-metadata-0" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.128304 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4\") " pod="openstack/nova-scheduler-0" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.128363 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4-config-data\") pod \"nova-scheduler-0\" (UID: \"e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4\") " pod="openstack/nova-scheduler-0" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.128481 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knchj\" (UniqueName: \"kubernetes.io/projected/e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4-kube-api-access-knchj\") pod \"nova-scheduler-0\" (UID: \"e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4\") " pod="openstack/nova-scheduler-0" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.141444 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4\") " pod="openstack/nova-scheduler-0" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.143003 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4-config-data\") pod \"nova-scheduler-0\" (UID: \"e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4\") " pod="openstack/nova-scheduler-0" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.151956 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d654f6767-wtpqm"] Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.153921 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d654f6767-wtpqm" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.161647 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knchj\" (UniqueName: \"kubernetes.io/projected/e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4-kube-api-access-knchj\") pod \"nova-scheduler-0\" (UID: \"e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4\") " pod="openstack/nova-scheduler-0" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.165019 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d654f6767-wtpqm"] Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.230534 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpmqh\" (UniqueName: \"kubernetes.io/projected/e9574c0c-29bf-43bc-b090-35ece3a54e94-kube-api-access-hpmqh\") pod \"dnsmasq-dns-5d654f6767-wtpqm\" (UID: \"e9574c0c-29bf-43bc-b090-35ece3a54e94\") " pod="openstack/dnsmasq-dns-5d654f6767-wtpqm" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.230604 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e9574c0c-29bf-43bc-b090-35ece3a54e94-ovsdbserver-nb\") pod \"dnsmasq-dns-5d654f6767-wtpqm\" (UID: \"e9574c0c-29bf-43bc-b090-35ece3a54e94\") " pod="openstack/dnsmasq-dns-5d654f6767-wtpqm" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.230628 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e9574c0c-29bf-43bc-b090-35ece3a54e94-ovsdbserver-sb\") pod \"dnsmasq-dns-5d654f6767-wtpqm\" (UID: \"e9574c0c-29bf-43bc-b090-35ece3a54e94\") " pod="openstack/dnsmasq-dns-5d654f6767-wtpqm" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.230678 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9574c0c-29bf-43bc-b090-35ece3a54e94-config\") pod \"dnsmasq-dns-5d654f6767-wtpqm\" (UID: \"e9574c0c-29bf-43bc-b090-35ece3a54e94\") " pod="openstack/dnsmasq-dns-5d654f6767-wtpqm" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.230706 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e9574c0c-29bf-43bc-b090-35ece3a54e94-dns-swift-storage-0\") pod \"dnsmasq-dns-5d654f6767-wtpqm\" (UID: \"e9574c0c-29bf-43bc-b090-35ece3a54e94\") " pod="openstack/dnsmasq-dns-5d654f6767-wtpqm" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.230767 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e9574c0c-29bf-43bc-b090-35ece3a54e94-dns-svc\") pod \"dnsmasq-dns-5d654f6767-wtpqm\" (UID: \"e9574c0c-29bf-43bc-b090-35ece3a54e94\") " pod="openstack/dnsmasq-dns-5d654f6767-wtpqm" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.294925 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.312554 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.335610 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpmqh\" (UniqueName: \"kubernetes.io/projected/e9574c0c-29bf-43bc-b090-35ece3a54e94-kube-api-access-hpmqh\") pod \"dnsmasq-dns-5d654f6767-wtpqm\" (UID: \"e9574c0c-29bf-43bc-b090-35ece3a54e94\") " pod="openstack/dnsmasq-dns-5d654f6767-wtpqm" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.335681 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e9574c0c-29bf-43bc-b090-35ece3a54e94-ovsdbserver-nb\") pod \"dnsmasq-dns-5d654f6767-wtpqm\" (UID: \"e9574c0c-29bf-43bc-b090-35ece3a54e94\") " pod="openstack/dnsmasq-dns-5d654f6767-wtpqm" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.335714 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e9574c0c-29bf-43bc-b090-35ece3a54e94-ovsdbserver-sb\") pod \"dnsmasq-dns-5d654f6767-wtpqm\" (UID: \"e9574c0c-29bf-43bc-b090-35ece3a54e94\") " pod="openstack/dnsmasq-dns-5d654f6767-wtpqm" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.335746 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9574c0c-29bf-43bc-b090-35ece3a54e94-config\") pod \"dnsmasq-dns-5d654f6767-wtpqm\" (UID: \"e9574c0c-29bf-43bc-b090-35ece3a54e94\") " pod="openstack/dnsmasq-dns-5d654f6767-wtpqm" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.335773 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e9574c0c-29bf-43bc-b090-35ece3a54e94-dns-swift-storage-0\") pod \"dnsmasq-dns-5d654f6767-wtpqm\" (UID: \"e9574c0c-29bf-43bc-b090-35ece3a54e94\") " pod="openstack/dnsmasq-dns-5d654f6767-wtpqm" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.335830 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e9574c0c-29bf-43bc-b090-35ece3a54e94-dns-svc\") pod \"dnsmasq-dns-5d654f6767-wtpqm\" (UID: \"e9574c0c-29bf-43bc-b090-35ece3a54e94\") " pod="openstack/dnsmasq-dns-5d654f6767-wtpqm" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.337554 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e9574c0c-29bf-43bc-b090-35ece3a54e94-dns-svc\") pod \"dnsmasq-dns-5d654f6767-wtpqm\" (UID: \"e9574c0c-29bf-43bc-b090-35ece3a54e94\") " pod="openstack/dnsmasq-dns-5d654f6767-wtpqm" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.338826 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e9574c0c-29bf-43bc-b090-35ece3a54e94-ovsdbserver-nb\") pod \"dnsmasq-dns-5d654f6767-wtpqm\" (UID: \"e9574c0c-29bf-43bc-b090-35ece3a54e94\") " pod="openstack/dnsmasq-dns-5d654f6767-wtpqm" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.340288 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e9574c0c-29bf-43bc-b090-35ece3a54e94-ovsdbserver-sb\") pod \"dnsmasq-dns-5d654f6767-wtpqm\" (UID: \"e9574c0c-29bf-43bc-b090-35ece3a54e94\") " pod="openstack/dnsmasq-dns-5d654f6767-wtpqm" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.342409 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9574c0c-29bf-43bc-b090-35ece3a54e94-config\") pod \"dnsmasq-dns-5d654f6767-wtpqm\" (UID: \"e9574c0c-29bf-43bc-b090-35ece3a54e94\") " pod="openstack/dnsmasq-dns-5d654f6767-wtpqm" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.346580 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e9574c0c-29bf-43bc-b090-35ece3a54e94-dns-swift-storage-0\") pod \"dnsmasq-dns-5d654f6767-wtpqm\" (UID: \"e9574c0c-29bf-43bc-b090-35ece3a54e94\") " pod="openstack/dnsmasq-dns-5d654f6767-wtpqm" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.358680 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpmqh\" (UniqueName: \"kubernetes.io/projected/e9574c0c-29bf-43bc-b090-35ece3a54e94-kube-api-access-hpmqh\") pod \"dnsmasq-dns-5d654f6767-wtpqm\" (UID: \"e9574c0c-29bf-43bc-b090-35ece3a54e94\") " pod="openstack/dnsmasq-dns-5d654f6767-wtpqm" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.494638 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d654f6767-wtpqm" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.609974 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-xglhv"] Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.784065 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 09:54:17 crc kubenswrapper[4787]: W1001 09:54:17.817990 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod06718a12_c4cb_43ed_be5f_0bee09272996.slice/crio-b8527834f98426a9d4182b1b89990c2c65334ca925a0ce00e7503df61992661f WatchSource:0}: Error finding container b8527834f98426a9d4182b1b89990c2c65334ca925a0ce00e7503df61992661f: Status 404 returned error can't find the container with id b8527834f98426a9d4182b1b89990c2c65334ca925a0ce00e7503df61992661f Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.826911 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.971793 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-jqjq6"] Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.973317 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-jqjq6" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.975889 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.982399 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 01 09:54:17 crc kubenswrapper[4787]: I1001 09:54:17.989878 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-jqjq6"] Oct 01 09:54:18 crc kubenswrapper[4787]: I1001 09:54:18.035261 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:54:18 crc kubenswrapper[4787]: W1001 09:54:18.039847 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c7de75f_5388_4bda_97ef_d3a17234001b.slice/crio-2a93b0ceec1a8b3b37ff1f36dee2a606d62812014f2e640244b35aca3d3d07b9 WatchSource:0}: Error finding container 2a93b0ceec1a8b3b37ff1f36dee2a606d62812014f2e640244b35aca3d3d07b9: Status 404 returned error can't find the container with id 2a93b0ceec1a8b3b37ff1f36dee2a606d62812014f2e640244b35aca3d3d07b9 Oct 01 09:54:18 crc kubenswrapper[4787]: I1001 09:54:18.109824 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bmtp\" (UniqueName: \"kubernetes.io/projected/a6ac40b1-e2a4-4d15-9855-6bbf9daad643-kube-api-access-8bmtp\") pod \"nova-cell1-conductor-db-sync-jqjq6\" (UID: \"a6ac40b1-e2a4-4d15-9855-6bbf9daad643\") " pod="openstack/nova-cell1-conductor-db-sync-jqjq6" Oct 01 09:54:18 crc kubenswrapper[4787]: I1001 09:54:18.110210 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6ac40b1-e2a4-4d15-9855-6bbf9daad643-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-jqjq6\" (UID: \"a6ac40b1-e2a4-4d15-9855-6bbf9daad643\") " pod="openstack/nova-cell1-conductor-db-sync-jqjq6" Oct 01 09:54:18 crc kubenswrapper[4787]: I1001 09:54:18.110258 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6ac40b1-e2a4-4d15-9855-6bbf9daad643-scripts\") pod \"nova-cell1-conductor-db-sync-jqjq6\" (UID: \"a6ac40b1-e2a4-4d15-9855-6bbf9daad643\") " pod="openstack/nova-cell1-conductor-db-sync-jqjq6" Oct 01 09:54:18 crc kubenswrapper[4787]: I1001 09:54:18.110346 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6ac40b1-e2a4-4d15-9855-6bbf9daad643-config-data\") pod \"nova-cell1-conductor-db-sync-jqjq6\" (UID: \"a6ac40b1-e2a4-4d15-9855-6bbf9daad643\") " pod="openstack/nova-cell1-conductor-db-sync-jqjq6" Oct 01 09:54:18 crc kubenswrapper[4787]: I1001 09:54:18.162380 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 09:54:18 crc kubenswrapper[4787]: I1001 09:54:18.211726 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6ac40b1-e2a4-4d15-9855-6bbf9daad643-config-data\") pod \"nova-cell1-conductor-db-sync-jqjq6\" (UID: \"a6ac40b1-e2a4-4d15-9855-6bbf9daad643\") " pod="openstack/nova-cell1-conductor-db-sync-jqjq6" Oct 01 09:54:18 crc kubenswrapper[4787]: I1001 09:54:18.211820 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bmtp\" (UniqueName: \"kubernetes.io/projected/a6ac40b1-e2a4-4d15-9855-6bbf9daad643-kube-api-access-8bmtp\") pod \"nova-cell1-conductor-db-sync-jqjq6\" (UID: \"a6ac40b1-e2a4-4d15-9855-6bbf9daad643\") " pod="openstack/nova-cell1-conductor-db-sync-jqjq6" Oct 01 09:54:18 crc kubenswrapper[4787]: I1001 09:54:18.211887 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6ac40b1-e2a4-4d15-9855-6bbf9daad643-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-jqjq6\" (UID: \"a6ac40b1-e2a4-4d15-9855-6bbf9daad643\") " pod="openstack/nova-cell1-conductor-db-sync-jqjq6" Oct 01 09:54:18 crc kubenswrapper[4787]: I1001 09:54:18.212852 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6ac40b1-e2a4-4d15-9855-6bbf9daad643-scripts\") pod \"nova-cell1-conductor-db-sync-jqjq6\" (UID: \"a6ac40b1-e2a4-4d15-9855-6bbf9daad643\") " pod="openstack/nova-cell1-conductor-db-sync-jqjq6" Oct 01 09:54:18 crc kubenswrapper[4787]: I1001 09:54:18.223436 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6ac40b1-e2a4-4d15-9855-6bbf9daad643-config-data\") pod \"nova-cell1-conductor-db-sync-jqjq6\" (UID: \"a6ac40b1-e2a4-4d15-9855-6bbf9daad643\") " pod="openstack/nova-cell1-conductor-db-sync-jqjq6" Oct 01 09:54:18 crc kubenswrapper[4787]: I1001 09:54:18.223706 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6ac40b1-e2a4-4d15-9855-6bbf9daad643-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-jqjq6\" (UID: \"a6ac40b1-e2a4-4d15-9855-6bbf9daad643\") " pod="openstack/nova-cell1-conductor-db-sync-jqjq6" Oct 01 09:54:18 crc kubenswrapper[4787]: I1001 09:54:18.224188 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6ac40b1-e2a4-4d15-9855-6bbf9daad643-scripts\") pod \"nova-cell1-conductor-db-sync-jqjq6\" (UID: \"a6ac40b1-e2a4-4d15-9855-6bbf9daad643\") " pod="openstack/nova-cell1-conductor-db-sync-jqjq6" Oct 01 09:54:18 crc kubenswrapper[4787]: I1001 09:54:18.238229 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bmtp\" (UniqueName: \"kubernetes.io/projected/a6ac40b1-e2a4-4d15-9855-6bbf9daad643-kube-api-access-8bmtp\") pod \"nova-cell1-conductor-db-sync-jqjq6\" (UID: \"a6ac40b1-e2a4-4d15-9855-6bbf9daad643\") " pod="openstack/nova-cell1-conductor-db-sync-jqjq6" Oct 01 09:54:18 crc kubenswrapper[4787]: I1001 09:54:18.288817 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d654f6767-wtpqm"] Oct 01 09:54:18 crc kubenswrapper[4787]: I1001 09:54:18.310387 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-jqjq6" Oct 01 09:54:18 crc kubenswrapper[4787]: I1001 09:54:18.483247 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"13182f8e-43d0-40f7-b5a4-f11f34998d3d","Type":"ContainerStarted","Data":"2bd00980e99795af90e8253e592241aaeb05bf08b929df0b7c2bdf43b10afc37"} Oct 01 09:54:18 crc kubenswrapper[4787]: I1001 09:54:18.484596 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d654f6767-wtpqm" event={"ID":"e9574c0c-29bf-43bc-b090-35ece3a54e94","Type":"ContainerStarted","Data":"110b081b814c531a86a8e3a9b5b7bc79dde6b5ced503abe1f979ebf41462ebdd"} Oct 01 09:54:18 crc kubenswrapper[4787]: I1001 09:54:18.485977 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4","Type":"ContainerStarted","Data":"03eae54e5348fb9b7eb50caa9a87b7f97d89d1dbab696ea5117dbb0f5f73f7ce"} Oct 01 09:54:18 crc kubenswrapper[4787]: I1001 09:54:18.487478 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"06718a12-c4cb-43ed-be5f-0bee09272996","Type":"ContainerStarted","Data":"b8527834f98426a9d4182b1b89990c2c65334ca925a0ce00e7503df61992661f"} Oct 01 09:54:18 crc kubenswrapper[4787]: I1001 09:54:18.488408 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2c7de75f-5388-4bda-97ef-d3a17234001b","Type":"ContainerStarted","Data":"2a93b0ceec1a8b3b37ff1f36dee2a606d62812014f2e640244b35aca3d3d07b9"} Oct 01 09:54:18 crc kubenswrapper[4787]: I1001 09:54:18.490346 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-xglhv" event={"ID":"0634a251-9095-454e-8a37-ff57c2ec40e1","Type":"ContainerStarted","Data":"515b5403724fa4d69f5a7ff8da7ef602083269aeb1885f1f9803483e65cb522d"} Oct 01 09:54:18 crc kubenswrapper[4787]: I1001 09:54:18.490366 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-xglhv" event={"ID":"0634a251-9095-454e-8a37-ff57c2ec40e1","Type":"ContainerStarted","Data":"e651567ed1aa0072c4add82685bca15d0654ccba32784ca6ec6c3a16ad373e88"} Oct 01 09:54:18 crc kubenswrapper[4787]: I1001 09:54:18.518631 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-xglhv" podStartSLOduration=2.518608265 podStartE2EDuration="2.518608265s" podCreationTimestamp="2025-10-01 09:54:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:54:18.516901992 +0000 UTC m=+1090.632046149" watchObservedRunningTime="2025-10-01 09:54:18.518608265 +0000 UTC m=+1090.633752422" Oct 01 09:54:18 crc kubenswrapper[4787]: I1001 09:54:18.905641 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-jqjq6"] Oct 01 09:54:19 crc kubenswrapper[4787]: I1001 09:54:19.521220 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-jqjq6" event={"ID":"a6ac40b1-e2a4-4d15-9855-6bbf9daad643","Type":"ContainerStarted","Data":"f1b49a0aadf27cc2a040d9304a795968e93c53ef38c082785703c4bf0b203b31"} Oct 01 09:54:19 crc kubenswrapper[4787]: I1001 09:54:19.521289 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-jqjq6" event={"ID":"a6ac40b1-e2a4-4d15-9855-6bbf9daad643","Type":"ContainerStarted","Data":"4fddbb9eee21cc194312748e661814ed6937ff29bc3430856ebbfd2318e643f5"} Oct 01 09:54:19 crc kubenswrapper[4787]: I1001 09:54:19.526630 4787 generic.go:334] "Generic (PLEG): container finished" podID="e9574c0c-29bf-43bc-b090-35ece3a54e94" containerID="cd830538ea3ad81ccaddbe633be05f3709972cac61e3d9ad12bb55b9a13bf772" exitCode=0 Oct 01 09:54:19 crc kubenswrapper[4787]: I1001 09:54:19.526698 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d654f6767-wtpqm" event={"ID":"e9574c0c-29bf-43bc-b090-35ece3a54e94","Type":"ContainerDied","Data":"cd830538ea3ad81ccaddbe633be05f3709972cac61e3d9ad12bb55b9a13bf772"} Oct 01 09:54:19 crc kubenswrapper[4787]: I1001 09:54:19.538807 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-jqjq6" podStartSLOduration=2.538785228 podStartE2EDuration="2.538785228s" podCreationTimestamp="2025-10-01 09:54:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:54:19.536218882 +0000 UTC m=+1091.651363059" watchObservedRunningTime="2025-10-01 09:54:19.538785228 +0000 UTC m=+1091.653929385" Oct 01 09:54:20 crc kubenswrapper[4787]: I1001 09:54:20.764168 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:54:20 crc kubenswrapper[4787]: I1001 09:54:20.798328 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 09:54:22 crc kubenswrapper[4787]: I1001 09:54:22.590955 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"13182f8e-43d0-40f7-b5a4-f11f34998d3d","Type":"ContainerStarted","Data":"e38fa43fb47cc1a3b63c611be02241530a0bcf30baada4f0f1151c07e73f71eb"} Oct 01 09:54:22 crc kubenswrapper[4787]: I1001 09:54:22.591799 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="13182f8e-43d0-40f7-b5a4-f11f34998d3d" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://e38fa43fb47cc1a3b63c611be02241530a0bcf30baada4f0f1151c07e73f71eb" gracePeriod=30 Oct 01 09:54:22 crc kubenswrapper[4787]: I1001 09:54:22.615687 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d654f6767-wtpqm" event={"ID":"e9574c0c-29bf-43bc-b090-35ece3a54e94","Type":"ContainerStarted","Data":"2e0b5dd5073c0fa32a8d6df15f44078a84e399426c318ffd3e06018034b44a89"} Oct 01 09:54:22 crc kubenswrapper[4787]: I1001 09:54:22.617965 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5d654f6767-wtpqm" Oct 01 09:54:22 crc kubenswrapper[4787]: I1001 09:54:22.635032 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.679000556 podStartE2EDuration="6.635008877s" podCreationTimestamp="2025-10-01 09:54:16 +0000 UTC" firstStartedPulling="2025-10-01 09:54:17.854893859 +0000 UTC m=+1089.970038016" lastFinishedPulling="2025-10-01 09:54:21.81090217 +0000 UTC m=+1093.926046337" observedRunningTime="2025-10-01 09:54:22.628737898 +0000 UTC m=+1094.743882055" watchObservedRunningTime="2025-10-01 09:54:22.635008877 +0000 UTC m=+1094.750153034" Oct 01 09:54:22 crc kubenswrapper[4787]: I1001 09:54:22.636051 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4","Type":"ContainerStarted","Data":"ab4386eec86d639ba0ad81456e5b1c2e571489777256838e3826cf0c99140712"} Oct 01 09:54:22 crc kubenswrapper[4787]: I1001 09:54:22.640016 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"06718a12-c4cb-43ed-be5f-0bee09272996","Type":"ContainerStarted","Data":"631ecef0955d0e36d72f5e7b8ed2bd3e36b84ae18a10e48d9c076aa7c2987b74"} Oct 01 09:54:22 crc kubenswrapper[4787]: I1001 09:54:22.646500 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2c7de75f-5388-4bda-97ef-d3a17234001b","Type":"ContainerStarted","Data":"fbcdcd11b746a170ee3eb44ccf26a77e4bc1c3687c06637d776d97b81b8f455b"} Oct 01 09:54:22 crc kubenswrapper[4787]: I1001 09:54:22.646820 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2c7de75f-5388-4bda-97ef-d3a17234001b" containerName="nova-metadata-log" containerID="cri-o://fbcdcd11b746a170ee3eb44ccf26a77e4bc1c3687c06637d776d97b81b8f455b" gracePeriod=30 Oct 01 09:54:22 crc kubenswrapper[4787]: I1001 09:54:22.647282 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2c7de75f-5388-4bda-97ef-d3a17234001b" containerName="nova-metadata-metadata" containerID="cri-o://50854a5e8554c38855622f13e3e07ceedacfd52a6adccf1ebbe564a1a31d2099" gracePeriod=30 Oct 01 09:54:22 crc kubenswrapper[4787]: I1001 09:54:22.663848 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5d654f6767-wtpqm" podStartSLOduration=5.663823937 podStartE2EDuration="5.663823937s" podCreationTimestamp="2025-10-01 09:54:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:54:22.657186429 +0000 UTC m=+1094.772330606" watchObservedRunningTime="2025-10-01 09:54:22.663823937 +0000 UTC m=+1094.778968094" Oct 01 09:54:22 crc kubenswrapper[4787]: I1001 09:54:22.706094 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.061514833 podStartE2EDuration="6.706040516s" podCreationTimestamp="2025-10-01 09:54:16 +0000 UTC" firstStartedPulling="2025-10-01 09:54:18.166258364 +0000 UTC m=+1090.281402521" lastFinishedPulling="2025-10-01 09:54:21.810784047 +0000 UTC m=+1093.925928204" observedRunningTime="2025-10-01 09:54:22.703633014 +0000 UTC m=+1094.818777181" watchObservedRunningTime="2025-10-01 09:54:22.706040516 +0000 UTC m=+1094.821184673" Oct 01 09:54:22 crc kubenswrapper[4787]: I1001 09:54:22.711018 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.946693425 podStartE2EDuration="6.711004572s" podCreationTimestamp="2025-10-01 09:54:16 +0000 UTC" firstStartedPulling="2025-10-01 09:54:18.044957582 +0000 UTC m=+1090.160101739" lastFinishedPulling="2025-10-01 09:54:21.809268729 +0000 UTC m=+1093.924412886" observedRunningTime="2025-10-01 09:54:22.680878969 +0000 UTC m=+1094.796023146" watchObservedRunningTime="2025-10-01 09:54:22.711004572 +0000 UTC m=+1094.826148729" Oct 01 09:54:23 crc kubenswrapper[4787]: I1001 09:54:23.658688 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"06718a12-c4cb-43ed-be5f-0bee09272996","Type":"ContainerStarted","Data":"72766222a72bbb3bc9a7b0d2308ee4218d433bed36ef7f8e8d0817cacabea125"} Oct 01 09:54:23 crc kubenswrapper[4787]: I1001 09:54:23.661782 4787 generic.go:334] "Generic (PLEG): container finished" podID="2c7de75f-5388-4bda-97ef-d3a17234001b" containerID="fbcdcd11b746a170ee3eb44ccf26a77e4bc1c3687c06637d776d97b81b8f455b" exitCode=143 Oct 01 09:54:23 crc kubenswrapper[4787]: I1001 09:54:23.661883 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2c7de75f-5388-4bda-97ef-d3a17234001b","Type":"ContainerDied","Data":"fbcdcd11b746a170ee3eb44ccf26a77e4bc1c3687c06637d776d97b81b8f455b"} Oct 01 09:54:23 crc kubenswrapper[4787]: I1001 09:54:23.661971 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2c7de75f-5388-4bda-97ef-d3a17234001b","Type":"ContainerStarted","Data":"50854a5e8554c38855622f13e3e07ceedacfd52a6adccf1ebbe564a1a31d2099"} Oct 01 09:54:23 crc kubenswrapper[4787]: I1001 09:54:23.683144 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.708490914 podStartE2EDuration="7.683125116s" podCreationTimestamp="2025-10-01 09:54:16 +0000 UTC" firstStartedPulling="2025-10-01 09:54:17.828226744 +0000 UTC m=+1089.943370901" lastFinishedPulling="2025-10-01 09:54:21.802860946 +0000 UTC m=+1093.918005103" observedRunningTime="2025-10-01 09:54:23.6817102 +0000 UTC m=+1095.796854367" watchObservedRunningTime="2025-10-01 09:54:23.683125116 +0000 UTC m=+1095.798269273" Oct 01 09:54:26 crc kubenswrapper[4787]: I1001 09:54:26.908856 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:54:26 crc kubenswrapper[4787]: I1001 09:54:26.972056 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 01 09:54:26 crc kubenswrapper[4787]: I1001 09:54:26.972180 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 01 09:54:27 crc kubenswrapper[4787]: I1001 09:54:27.296841 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 01 09:54:27 crc kubenswrapper[4787]: I1001 09:54:27.297460 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 01 09:54:27 crc kubenswrapper[4787]: I1001 09:54:27.314139 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 01 09:54:27 crc kubenswrapper[4787]: I1001 09:54:27.314257 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 01 09:54:27 crc kubenswrapper[4787]: I1001 09:54:27.367884 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 01 09:54:27 crc kubenswrapper[4787]: I1001 09:54:27.497722 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5d654f6767-wtpqm" Oct 01 09:54:27 crc kubenswrapper[4787]: I1001 09:54:27.612399 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bd7989875-t7k69"] Oct 01 09:54:27 crc kubenswrapper[4787]: I1001 09:54:27.613188 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7bd7989875-t7k69" podUID="951b7125-da65-4cfa-ae70-bd8013441bbb" containerName="dnsmasq-dns" containerID="cri-o://fabf0f56b0adabd51f151a10a9e41c01c0cfce1054514c925a914a256348f858" gracePeriod=10 Oct 01 09:54:27 crc kubenswrapper[4787]: I1001 09:54:27.713286 4787 generic.go:334] "Generic (PLEG): container finished" podID="a6ac40b1-e2a4-4d15-9855-6bbf9daad643" containerID="f1b49a0aadf27cc2a040d9304a795968e93c53ef38c082785703c4bf0b203b31" exitCode=0 Oct 01 09:54:27 crc kubenswrapper[4787]: I1001 09:54:27.713458 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-jqjq6" event={"ID":"a6ac40b1-e2a4-4d15-9855-6bbf9daad643","Type":"ContainerDied","Data":"f1b49a0aadf27cc2a040d9304a795968e93c53ef38c082785703c4bf0b203b31"} Oct 01 09:54:27 crc kubenswrapper[4787]: I1001 09:54:27.716150 4787 generic.go:334] "Generic (PLEG): container finished" podID="0634a251-9095-454e-8a37-ff57c2ec40e1" containerID="515b5403724fa4d69f5a7ff8da7ef602083269aeb1885f1f9803483e65cb522d" exitCode=0 Oct 01 09:54:27 crc kubenswrapper[4787]: I1001 09:54:27.716605 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-xglhv" event={"ID":"0634a251-9095-454e-8a37-ff57c2ec40e1","Type":"ContainerDied","Data":"515b5403724fa4d69f5a7ff8da7ef602083269aeb1885f1f9803483e65cb522d"} Oct 01 09:54:27 crc kubenswrapper[4787]: I1001 09:54:27.766704 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 01 09:54:28 crc kubenswrapper[4787]: I1001 09:54:28.070339 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="06718a12-c4cb-43ed-be5f-0bee09272996" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.189:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 01 09:54:28 crc kubenswrapper[4787]: I1001 09:54:28.070795 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="06718a12-c4cb-43ed-be5f-0bee09272996" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.189:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 01 09:54:28 crc kubenswrapper[4787]: I1001 09:54:28.320880 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bd7989875-t7k69" Oct 01 09:54:28 crc kubenswrapper[4787]: I1001 09:54:28.508594 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/951b7125-da65-4cfa-ae70-bd8013441bbb-ovsdbserver-sb\") pod \"951b7125-da65-4cfa-ae70-bd8013441bbb\" (UID: \"951b7125-da65-4cfa-ae70-bd8013441bbb\") " Oct 01 09:54:28 crc kubenswrapper[4787]: I1001 09:54:28.508699 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4bpb\" (UniqueName: \"kubernetes.io/projected/951b7125-da65-4cfa-ae70-bd8013441bbb-kube-api-access-q4bpb\") pod \"951b7125-da65-4cfa-ae70-bd8013441bbb\" (UID: \"951b7125-da65-4cfa-ae70-bd8013441bbb\") " Oct 01 09:54:28 crc kubenswrapper[4787]: I1001 09:54:28.508767 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/951b7125-da65-4cfa-ae70-bd8013441bbb-ovsdbserver-nb\") pod \"951b7125-da65-4cfa-ae70-bd8013441bbb\" (UID: \"951b7125-da65-4cfa-ae70-bd8013441bbb\") " Oct 01 09:54:28 crc kubenswrapper[4787]: I1001 09:54:28.508901 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/951b7125-da65-4cfa-ae70-bd8013441bbb-dns-swift-storage-0\") pod \"951b7125-da65-4cfa-ae70-bd8013441bbb\" (UID: \"951b7125-da65-4cfa-ae70-bd8013441bbb\") " Oct 01 09:54:28 crc kubenswrapper[4787]: I1001 09:54:28.509025 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/951b7125-da65-4cfa-ae70-bd8013441bbb-config\") pod \"951b7125-da65-4cfa-ae70-bd8013441bbb\" (UID: \"951b7125-da65-4cfa-ae70-bd8013441bbb\") " Oct 01 09:54:28 crc kubenswrapper[4787]: I1001 09:54:28.509042 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/951b7125-da65-4cfa-ae70-bd8013441bbb-dns-svc\") pod \"951b7125-da65-4cfa-ae70-bd8013441bbb\" (UID: \"951b7125-da65-4cfa-ae70-bd8013441bbb\") " Oct 01 09:54:28 crc kubenswrapper[4787]: I1001 09:54:28.531541 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/951b7125-da65-4cfa-ae70-bd8013441bbb-kube-api-access-q4bpb" (OuterVolumeSpecName: "kube-api-access-q4bpb") pod "951b7125-da65-4cfa-ae70-bd8013441bbb" (UID: "951b7125-da65-4cfa-ae70-bd8013441bbb"). InnerVolumeSpecName "kube-api-access-q4bpb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:54:28 crc kubenswrapper[4787]: I1001 09:54:28.580477 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/951b7125-da65-4cfa-ae70-bd8013441bbb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "951b7125-da65-4cfa-ae70-bd8013441bbb" (UID: "951b7125-da65-4cfa-ae70-bd8013441bbb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:54:28 crc kubenswrapper[4787]: I1001 09:54:28.599615 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/951b7125-da65-4cfa-ae70-bd8013441bbb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "951b7125-da65-4cfa-ae70-bd8013441bbb" (UID: "951b7125-da65-4cfa-ae70-bd8013441bbb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:54:28 crc kubenswrapper[4787]: I1001 09:54:28.602736 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/951b7125-da65-4cfa-ae70-bd8013441bbb-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "951b7125-da65-4cfa-ae70-bd8013441bbb" (UID: "951b7125-da65-4cfa-ae70-bd8013441bbb"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:54:28 crc kubenswrapper[4787]: I1001 09:54:28.603788 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/951b7125-da65-4cfa-ae70-bd8013441bbb-config" (OuterVolumeSpecName: "config") pod "951b7125-da65-4cfa-ae70-bd8013441bbb" (UID: "951b7125-da65-4cfa-ae70-bd8013441bbb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:54:28 crc kubenswrapper[4787]: I1001 09:54:28.611206 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/951b7125-da65-4cfa-ae70-bd8013441bbb-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:28 crc kubenswrapper[4787]: I1001 09:54:28.611246 4787 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/951b7125-da65-4cfa-ae70-bd8013441bbb-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:28 crc kubenswrapper[4787]: I1001 09:54:28.611258 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4bpb\" (UniqueName: \"kubernetes.io/projected/951b7125-da65-4cfa-ae70-bd8013441bbb-kube-api-access-q4bpb\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:28 crc kubenswrapper[4787]: I1001 09:54:28.611271 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/951b7125-da65-4cfa-ae70-bd8013441bbb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:28 crc kubenswrapper[4787]: I1001 09:54:28.611661 4787 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/951b7125-da65-4cfa-ae70-bd8013441bbb-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:28 crc kubenswrapper[4787]: I1001 09:54:28.620363 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/951b7125-da65-4cfa-ae70-bd8013441bbb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "951b7125-da65-4cfa-ae70-bd8013441bbb" (UID: "951b7125-da65-4cfa-ae70-bd8013441bbb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:54:28 crc kubenswrapper[4787]: I1001 09:54:28.713900 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/951b7125-da65-4cfa-ae70-bd8013441bbb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:28 crc kubenswrapper[4787]: I1001 09:54:28.739530 4787 generic.go:334] "Generic (PLEG): container finished" podID="951b7125-da65-4cfa-ae70-bd8013441bbb" containerID="fabf0f56b0adabd51f151a10a9e41c01c0cfce1054514c925a914a256348f858" exitCode=0 Oct 01 09:54:28 crc kubenswrapper[4787]: I1001 09:54:28.739657 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bd7989875-t7k69" event={"ID":"951b7125-da65-4cfa-ae70-bd8013441bbb","Type":"ContainerDied","Data":"fabf0f56b0adabd51f151a10a9e41c01c0cfce1054514c925a914a256348f858"} Oct 01 09:54:28 crc kubenswrapper[4787]: I1001 09:54:28.739692 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bd7989875-t7k69" event={"ID":"951b7125-da65-4cfa-ae70-bd8013441bbb","Type":"ContainerDied","Data":"e2da0c0e33808d07d5aeec7061eac40ace897a5c666adde108505fad19e83eec"} Oct 01 09:54:28 crc kubenswrapper[4787]: I1001 09:54:28.739702 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bd7989875-t7k69" Oct 01 09:54:28 crc kubenswrapper[4787]: I1001 09:54:28.739716 4787 scope.go:117] "RemoveContainer" containerID="fabf0f56b0adabd51f151a10a9e41c01c0cfce1054514c925a914a256348f858" Oct 01 09:54:28 crc kubenswrapper[4787]: I1001 09:54:28.830318 4787 scope.go:117] "RemoveContainer" containerID="b55726dfcd0b4cf35b279f0b055e1b5ce6a6a33e7784cc6dba6458a9b93b74b7" Oct 01 09:54:28 crc kubenswrapper[4787]: I1001 09:54:28.835590 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bd7989875-t7k69"] Oct 01 09:54:28 crc kubenswrapper[4787]: I1001 09:54:28.843764 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7bd7989875-t7k69"] Oct 01 09:54:28 crc kubenswrapper[4787]: I1001 09:54:28.860457 4787 scope.go:117] "RemoveContainer" containerID="fabf0f56b0adabd51f151a10a9e41c01c0cfce1054514c925a914a256348f858" Oct 01 09:54:28 crc kubenswrapper[4787]: E1001 09:54:28.861731 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fabf0f56b0adabd51f151a10a9e41c01c0cfce1054514c925a914a256348f858\": container with ID starting with fabf0f56b0adabd51f151a10a9e41c01c0cfce1054514c925a914a256348f858 not found: ID does not exist" containerID="fabf0f56b0adabd51f151a10a9e41c01c0cfce1054514c925a914a256348f858" Oct 01 09:54:28 crc kubenswrapper[4787]: I1001 09:54:28.861767 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fabf0f56b0adabd51f151a10a9e41c01c0cfce1054514c925a914a256348f858"} err="failed to get container status \"fabf0f56b0adabd51f151a10a9e41c01c0cfce1054514c925a914a256348f858\": rpc error: code = NotFound desc = could not find container \"fabf0f56b0adabd51f151a10a9e41c01c0cfce1054514c925a914a256348f858\": container with ID starting with fabf0f56b0adabd51f151a10a9e41c01c0cfce1054514c925a914a256348f858 not found: ID does not exist" Oct 01 09:54:28 crc kubenswrapper[4787]: I1001 09:54:28.861792 4787 scope.go:117] "RemoveContainer" containerID="b55726dfcd0b4cf35b279f0b055e1b5ce6a6a33e7784cc6dba6458a9b93b74b7" Oct 01 09:54:28 crc kubenswrapper[4787]: E1001 09:54:28.862215 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b55726dfcd0b4cf35b279f0b055e1b5ce6a6a33e7784cc6dba6458a9b93b74b7\": container with ID starting with b55726dfcd0b4cf35b279f0b055e1b5ce6a6a33e7784cc6dba6458a9b93b74b7 not found: ID does not exist" containerID="b55726dfcd0b4cf35b279f0b055e1b5ce6a6a33e7784cc6dba6458a9b93b74b7" Oct 01 09:54:28 crc kubenswrapper[4787]: I1001 09:54:28.862252 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b55726dfcd0b4cf35b279f0b055e1b5ce6a6a33e7784cc6dba6458a9b93b74b7"} err="failed to get container status \"b55726dfcd0b4cf35b279f0b055e1b5ce6a6a33e7784cc6dba6458a9b93b74b7\": rpc error: code = NotFound desc = could not find container \"b55726dfcd0b4cf35b279f0b055e1b5ce6a6a33e7784cc6dba6458a9b93b74b7\": container with ID starting with b55726dfcd0b4cf35b279f0b055e1b5ce6a6a33e7784cc6dba6458a9b93b74b7 not found: ID does not exist" Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.304252 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-xglhv" Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.438365 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0634a251-9095-454e-8a37-ff57c2ec40e1-scripts\") pod \"0634a251-9095-454e-8a37-ff57c2ec40e1\" (UID: \"0634a251-9095-454e-8a37-ff57c2ec40e1\") " Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.438865 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgllc\" (UniqueName: \"kubernetes.io/projected/0634a251-9095-454e-8a37-ff57c2ec40e1-kube-api-access-zgllc\") pod \"0634a251-9095-454e-8a37-ff57c2ec40e1\" (UID: \"0634a251-9095-454e-8a37-ff57c2ec40e1\") " Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.438913 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0634a251-9095-454e-8a37-ff57c2ec40e1-config-data\") pod \"0634a251-9095-454e-8a37-ff57c2ec40e1\" (UID: \"0634a251-9095-454e-8a37-ff57c2ec40e1\") " Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.439004 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0634a251-9095-454e-8a37-ff57c2ec40e1-combined-ca-bundle\") pod \"0634a251-9095-454e-8a37-ff57c2ec40e1\" (UID: \"0634a251-9095-454e-8a37-ff57c2ec40e1\") " Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.442688 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-jqjq6" Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.447659 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0634a251-9095-454e-8a37-ff57c2ec40e1-kube-api-access-zgllc" (OuterVolumeSpecName: "kube-api-access-zgllc") pod "0634a251-9095-454e-8a37-ff57c2ec40e1" (UID: "0634a251-9095-454e-8a37-ff57c2ec40e1"). InnerVolumeSpecName "kube-api-access-zgllc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.450367 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0634a251-9095-454e-8a37-ff57c2ec40e1-scripts" (OuterVolumeSpecName: "scripts") pod "0634a251-9095-454e-8a37-ff57c2ec40e1" (UID: "0634a251-9095-454e-8a37-ff57c2ec40e1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.474535 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0634a251-9095-454e-8a37-ff57c2ec40e1-config-data" (OuterVolumeSpecName: "config-data") pod "0634a251-9095-454e-8a37-ff57c2ec40e1" (UID: "0634a251-9095-454e-8a37-ff57c2ec40e1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.504826 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0634a251-9095-454e-8a37-ff57c2ec40e1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0634a251-9095-454e-8a37-ff57c2ec40e1" (UID: "0634a251-9095-454e-8a37-ff57c2ec40e1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.541786 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6ac40b1-e2a4-4d15-9855-6bbf9daad643-config-data\") pod \"a6ac40b1-e2a4-4d15-9855-6bbf9daad643\" (UID: \"a6ac40b1-e2a4-4d15-9855-6bbf9daad643\") " Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.542018 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6ac40b1-e2a4-4d15-9855-6bbf9daad643-combined-ca-bundle\") pod \"a6ac40b1-e2a4-4d15-9855-6bbf9daad643\" (UID: \"a6ac40b1-e2a4-4d15-9855-6bbf9daad643\") " Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.542090 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8bmtp\" (UniqueName: \"kubernetes.io/projected/a6ac40b1-e2a4-4d15-9855-6bbf9daad643-kube-api-access-8bmtp\") pod \"a6ac40b1-e2a4-4d15-9855-6bbf9daad643\" (UID: \"a6ac40b1-e2a4-4d15-9855-6bbf9daad643\") " Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.542131 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6ac40b1-e2a4-4d15-9855-6bbf9daad643-scripts\") pod \"a6ac40b1-e2a4-4d15-9855-6bbf9daad643\" (UID: \"a6ac40b1-e2a4-4d15-9855-6bbf9daad643\") " Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.542653 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0634a251-9095-454e-8a37-ff57c2ec40e1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.542670 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0634a251-9095-454e-8a37-ff57c2ec40e1-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.542681 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgllc\" (UniqueName: \"kubernetes.io/projected/0634a251-9095-454e-8a37-ff57c2ec40e1-kube-api-access-zgllc\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.542693 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0634a251-9095-454e-8a37-ff57c2ec40e1-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.545925 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6ac40b1-e2a4-4d15-9855-6bbf9daad643-scripts" (OuterVolumeSpecName: "scripts") pod "a6ac40b1-e2a4-4d15-9855-6bbf9daad643" (UID: "a6ac40b1-e2a4-4d15-9855-6bbf9daad643"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.547756 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6ac40b1-e2a4-4d15-9855-6bbf9daad643-kube-api-access-8bmtp" (OuterVolumeSpecName: "kube-api-access-8bmtp") pod "a6ac40b1-e2a4-4d15-9855-6bbf9daad643" (UID: "a6ac40b1-e2a4-4d15-9855-6bbf9daad643"). InnerVolumeSpecName "kube-api-access-8bmtp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.571962 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6ac40b1-e2a4-4d15-9855-6bbf9daad643-config-data" (OuterVolumeSpecName: "config-data") pod "a6ac40b1-e2a4-4d15-9855-6bbf9daad643" (UID: "a6ac40b1-e2a4-4d15-9855-6bbf9daad643"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.593643 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6ac40b1-e2a4-4d15-9855-6bbf9daad643-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a6ac40b1-e2a4-4d15-9855-6bbf9daad643" (UID: "a6ac40b1-e2a4-4d15-9855-6bbf9daad643"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.644036 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6ac40b1-e2a4-4d15-9855-6bbf9daad643-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.644070 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8bmtp\" (UniqueName: \"kubernetes.io/projected/a6ac40b1-e2a4-4d15-9855-6bbf9daad643-kube-api-access-8bmtp\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.644104 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6ac40b1-e2a4-4d15-9855-6bbf9daad643-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.644113 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6ac40b1-e2a4-4d15-9855-6bbf9daad643-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.752007 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-jqjq6" event={"ID":"a6ac40b1-e2a4-4d15-9855-6bbf9daad643","Type":"ContainerDied","Data":"4fddbb9eee21cc194312748e661814ed6937ff29bc3430856ebbfd2318e643f5"} Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.752055 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4fddbb9eee21cc194312748e661814ed6937ff29bc3430856ebbfd2318e643f5" Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.752053 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-jqjq6" Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.757792 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-xglhv" event={"ID":"0634a251-9095-454e-8a37-ff57c2ec40e1","Type":"ContainerDied","Data":"e651567ed1aa0072c4add82685bca15d0654ccba32784ca6ec6c3a16ad373e88"} Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.757827 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e651567ed1aa0072c4add82685bca15d0654ccba32784ca6ec6c3a16ad373e88" Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.757899 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-xglhv" Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.905983 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 01 09:54:29 crc kubenswrapper[4787]: E1001 09:54:29.906677 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="951b7125-da65-4cfa-ae70-bd8013441bbb" containerName="dnsmasq-dns" Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.906706 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="951b7125-da65-4cfa-ae70-bd8013441bbb" containerName="dnsmasq-dns" Oct 01 09:54:29 crc kubenswrapper[4787]: E1001 09:54:29.906741 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6ac40b1-e2a4-4d15-9855-6bbf9daad643" containerName="nova-cell1-conductor-db-sync" Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.906749 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6ac40b1-e2a4-4d15-9855-6bbf9daad643" containerName="nova-cell1-conductor-db-sync" Oct 01 09:54:29 crc kubenswrapper[4787]: E1001 09:54:29.906764 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="951b7125-da65-4cfa-ae70-bd8013441bbb" containerName="init" Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.906771 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="951b7125-da65-4cfa-ae70-bd8013441bbb" containerName="init" Oct 01 09:54:29 crc kubenswrapper[4787]: E1001 09:54:29.906788 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0634a251-9095-454e-8a37-ff57c2ec40e1" containerName="nova-manage" Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.906795 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="0634a251-9095-454e-8a37-ff57c2ec40e1" containerName="nova-manage" Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.907013 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6ac40b1-e2a4-4d15-9855-6bbf9daad643" containerName="nova-cell1-conductor-db-sync" Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.907036 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="0634a251-9095-454e-8a37-ff57c2ec40e1" containerName="nova-manage" Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.907059 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="951b7125-da65-4cfa-ae70-bd8013441bbb" containerName="dnsmasq-dns" Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.907883 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.914476 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 01 09:54:29 crc kubenswrapper[4787]: I1001 09:54:29.919238 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 01 09:54:30 crc kubenswrapper[4787]: I1001 09:54:30.053203 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f4a94ee-d626-48f8-93c2-547212f23da4-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"1f4a94ee-d626-48f8-93c2-547212f23da4\") " pod="openstack/nova-cell1-conductor-0" Oct 01 09:54:30 crc kubenswrapper[4787]: I1001 09:54:30.053269 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f4a94ee-d626-48f8-93c2-547212f23da4-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"1f4a94ee-d626-48f8-93c2-547212f23da4\") " pod="openstack/nova-cell1-conductor-0" Oct 01 09:54:30 crc kubenswrapper[4787]: I1001 09:54:30.053311 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cndvn\" (UniqueName: \"kubernetes.io/projected/1f4a94ee-d626-48f8-93c2-547212f23da4-kube-api-access-cndvn\") pod \"nova-cell1-conductor-0\" (UID: \"1f4a94ee-d626-48f8-93c2-547212f23da4\") " pod="openstack/nova-cell1-conductor-0" Oct 01 09:54:30 crc kubenswrapper[4787]: I1001 09:54:30.069061 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 01 09:54:30 crc kubenswrapper[4787]: I1001 09:54:30.069376 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="06718a12-c4cb-43ed-be5f-0bee09272996" containerName="nova-api-log" containerID="cri-o://631ecef0955d0e36d72f5e7b8ed2bd3e36b84ae18a10e48d9c076aa7c2987b74" gracePeriod=30 Oct 01 09:54:30 crc kubenswrapper[4787]: I1001 09:54:30.069574 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="06718a12-c4cb-43ed-be5f-0bee09272996" containerName="nova-api-api" containerID="cri-o://72766222a72bbb3bc9a7b0d2308ee4218d433bed36ef7f8e8d0817cacabea125" gracePeriod=30 Oct 01 09:54:30 crc kubenswrapper[4787]: I1001 09:54:30.095663 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 09:54:30 crc kubenswrapper[4787]: I1001 09:54:30.096393 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4" containerName="nova-scheduler-scheduler" containerID="cri-o://ab4386eec86d639ba0ad81456e5b1c2e571489777256838e3826cf0c99140712" gracePeriod=30 Oct 01 09:54:30 crc kubenswrapper[4787]: I1001 09:54:30.156832 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f4a94ee-d626-48f8-93c2-547212f23da4-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"1f4a94ee-d626-48f8-93c2-547212f23da4\") " pod="openstack/nova-cell1-conductor-0" Oct 01 09:54:30 crc kubenswrapper[4787]: I1001 09:54:30.156909 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f4a94ee-d626-48f8-93c2-547212f23da4-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"1f4a94ee-d626-48f8-93c2-547212f23da4\") " pod="openstack/nova-cell1-conductor-0" Oct 01 09:54:30 crc kubenswrapper[4787]: I1001 09:54:30.156964 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cndvn\" (UniqueName: \"kubernetes.io/projected/1f4a94ee-d626-48f8-93c2-547212f23da4-kube-api-access-cndvn\") pod \"nova-cell1-conductor-0\" (UID: \"1f4a94ee-d626-48f8-93c2-547212f23da4\") " pod="openstack/nova-cell1-conductor-0" Oct 01 09:54:30 crc kubenswrapper[4787]: I1001 09:54:30.162428 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f4a94ee-d626-48f8-93c2-547212f23da4-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"1f4a94ee-d626-48f8-93c2-547212f23da4\") " pod="openstack/nova-cell1-conductor-0" Oct 01 09:54:30 crc kubenswrapper[4787]: I1001 09:54:30.162558 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f4a94ee-d626-48f8-93c2-547212f23da4-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"1f4a94ee-d626-48f8-93c2-547212f23da4\") " pod="openstack/nova-cell1-conductor-0" Oct 01 09:54:30 crc kubenswrapper[4787]: I1001 09:54:30.178968 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cndvn\" (UniqueName: \"kubernetes.io/projected/1f4a94ee-d626-48f8-93c2-547212f23da4-kube-api-access-cndvn\") pod \"nova-cell1-conductor-0\" (UID: \"1f4a94ee-d626-48f8-93c2-547212f23da4\") " pod="openstack/nova-cell1-conductor-0" Oct 01 09:54:30 crc kubenswrapper[4787]: I1001 09:54:30.230609 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 01 09:54:30 crc kubenswrapper[4787]: I1001 09:54:30.548535 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="951b7125-da65-4cfa-ae70-bd8013441bbb" path="/var/lib/kubelet/pods/951b7125-da65-4cfa-ae70-bd8013441bbb/volumes" Oct 01 09:54:30 crc kubenswrapper[4787]: I1001 09:54:30.740610 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 01 09:54:30 crc kubenswrapper[4787]: I1001 09:54:30.796756 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"1f4a94ee-d626-48f8-93c2-547212f23da4","Type":"ContainerStarted","Data":"e4bb51eaa360aa8755ed71938300a7f5905782df1037dbd944f48a432379d371"} Oct 01 09:54:30 crc kubenswrapper[4787]: I1001 09:54:30.810991 4787 generic.go:334] "Generic (PLEG): container finished" podID="06718a12-c4cb-43ed-be5f-0bee09272996" containerID="631ecef0955d0e36d72f5e7b8ed2bd3e36b84ae18a10e48d9c076aa7c2987b74" exitCode=143 Oct 01 09:54:30 crc kubenswrapper[4787]: I1001 09:54:30.811107 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"06718a12-c4cb-43ed-be5f-0bee09272996","Type":"ContainerDied","Data":"631ecef0955d0e36d72f5e7b8ed2bd3e36b84ae18a10e48d9c076aa7c2987b74"} Oct 01 09:54:31 crc kubenswrapper[4787]: I1001 09:54:31.825147 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"1f4a94ee-d626-48f8-93c2-547212f23da4","Type":"ContainerStarted","Data":"5fcd40944ae049402a470044b1db715224f3d4b1bce8ff1bdc68644309ce7382"} Oct 01 09:54:31 crc kubenswrapper[4787]: I1001 09:54:31.826329 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 01 09:54:32 crc kubenswrapper[4787]: E1001 09:54:32.315369 4787 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ab4386eec86d639ba0ad81456e5b1c2e571489777256838e3826cf0c99140712" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 01 09:54:32 crc kubenswrapper[4787]: E1001 09:54:32.316952 4787 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ab4386eec86d639ba0ad81456e5b1c2e571489777256838e3826cf0c99140712" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 01 09:54:32 crc kubenswrapper[4787]: E1001 09:54:32.319677 4787 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ab4386eec86d639ba0ad81456e5b1c2e571489777256838e3826cf0c99140712" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 01 09:54:32 crc kubenswrapper[4787]: E1001 09:54:32.319729 4787 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4" containerName="nova-scheduler-scheduler" Oct 01 09:54:33 crc kubenswrapper[4787]: I1001 09:54:33.267322 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7bd7989875-t7k69" podUID="951b7125-da65-4cfa-ae70-bd8013441bbb" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.172:5353: i/o timeout" Oct 01 09:54:33 crc kubenswrapper[4787]: I1001 09:54:33.803556 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 01 09:54:33 crc kubenswrapper[4787]: I1001 09:54:33.830737 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=4.830715007 podStartE2EDuration="4.830715007s" podCreationTimestamp="2025-10-01 09:54:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:54:31.850651884 +0000 UTC m=+1103.965796041" watchObservedRunningTime="2025-10-01 09:54:33.830715007 +0000 UTC m=+1105.945859164" Oct 01 09:54:33 crc kubenswrapper[4787]: I1001 09:54:33.862189 4787 generic.go:334] "Generic (PLEG): container finished" podID="06718a12-c4cb-43ed-be5f-0bee09272996" containerID="72766222a72bbb3bc9a7b0d2308ee4218d433bed36ef7f8e8d0817cacabea125" exitCode=0 Oct 01 09:54:33 crc kubenswrapper[4787]: I1001 09:54:33.862254 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"06718a12-c4cb-43ed-be5f-0bee09272996","Type":"ContainerDied","Data":"72766222a72bbb3bc9a7b0d2308ee4218d433bed36ef7f8e8d0817cacabea125"} Oct 01 09:54:34 crc kubenswrapper[4787]: I1001 09:54:34.000464 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 09:54:34 crc kubenswrapper[4787]: I1001 09:54:34.079156 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06718a12-c4cb-43ed-be5f-0bee09272996-config-data\") pod \"06718a12-c4cb-43ed-be5f-0bee09272996\" (UID: \"06718a12-c4cb-43ed-be5f-0bee09272996\") " Oct 01 09:54:34 crc kubenswrapper[4787]: I1001 09:54:34.079275 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhbgg\" (UniqueName: \"kubernetes.io/projected/06718a12-c4cb-43ed-be5f-0bee09272996-kube-api-access-zhbgg\") pod \"06718a12-c4cb-43ed-be5f-0bee09272996\" (UID: \"06718a12-c4cb-43ed-be5f-0bee09272996\") " Oct 01 09:54:34 crc kubenswrapper[4787]: I1001 09:54:34.079323 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/06718a12-c4cb-43ed-be5f-0bee09272996-logs\") pod \"06718a12-c4cb-43ed-be5f-0bee09272996\" (UID: \"06718a12-c4cb-43ed-be5f-0bee09272996\") " Oct 01 09:54:34 crc kubenswrapper[4787]: I1001 09:54:34.079426 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06718a12-c4cb-43ed-be5f-0bee09272996-combined-ca-bundle\") pod \"06718a12-c4cb-43ed-be5f-0bee09272996\" (UID: \"06718a12-c4cb-43ed-be5f-0bee09272996\") " Oct 01 09:54:34 crc kubenswrapper[4787]: I1001 09:54:34.080043 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06718a12-c4cb-43ed-be5f-0bee09272996-logs" (OuterVolumeSpecName: "logs") pod "06718a12-c4cb-43ed-be5f-0bee09272996" (UID: "06718a12-c4cb-43ed-be5f-0bee09272996"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:54:34 crc kubenswrapper[4787]: I1001 09:54:34.100200 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06718a12-c4cb-43ed-be5f-0bee09272996-kube-api-access-zhbgg" (OuterVolumeSpecName: "kube-api-access-zhbgg") pod "06718a12-c4cb-43ed-be5f-0bee09272996" (UID: "06718a12-c4cb-43ed-be5f-0bee09272996"). InnerVolumeSpecName "kube-api-access-zhbgg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:54:34 crc kubenswrapper[4787]: I1001 09:54:34.112193 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06718a12-c4cb-43ed-be5f-0bee09272996-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "06718a12-c4cb-43ed-be5f-0bee09272996" (UID: "06718a12-c4cb-43ed-be5f-0bee09272996"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:54:34 crc kubenswrapper[4787]: I1001 09:54:34.116151 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06718a12-c4cb-43ed-be5f-0bee09272996-config-data" (OuterVolumeSpecName: "config-data") pod "06718a12-c4cb-43ed-be5f-0bee09272996" (UID: "06718a12-c4cb-43ed-be5f-0bee09272996"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:54:34 crc kubenswrapper[4787]: I1001 09:54:34.181999 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06718a12-c4cb-43ed-be5f-0bee09272996-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:34 crc kubenswrapper[4787]: I1001 09:54:34.182037 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06718a12-c4cb-43ed-be5f-0bee09272996-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:34 crc kubenswrapper[4787]: I1001 09:54:34.182047 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhbgg\" (UniqueName: \"kubernetes.io/projected/06718a12-c4cb-43ed-be5f-0bee09272996-kube-api-access-zhbgg\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:34 crc kubenswrapper[4787]: I1001 09:54:34.182062 4787 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/06718a12-c4cb-43ed-be5f-0bee09272996-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:34 crc kubenswrapper[4787]: I1001 09:54:34.493675 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 09:54:34 crc kubenswrapper[4787]: I1001 09:54:34.587908 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4-combined-ca-bundle\") pod \"e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4\" (UID: \"e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4\") " Oct 01 09:54:34 crc kubenswrapper[4787]: I1001 09:54:34.588995 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4-config-data\") pod \"e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4\" (UID: \"e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4\") " Oct 01 09:54:34 crc kubenswrapper[4787]: I1001 09:54:34.589217 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-knchj\" (UniqueName: \"kubernetes.io/projected/e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4-kube-api-access-knchj\") pod \"e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4\" (UID: \"e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4\") " Oct 01 09:54:34 crc kubenswrapper[4787]: I1001 09:54:34.593528 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4-kube-api-access-knchj" (OuterVolumeSpecName: "kube-api-access-knchj") pod "e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4" (UID: "e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4"). InnerVolumeSpecName "kube-api-access-knchj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:54:34 crc kubenswrapper[4787]: I1001 09:54:34.626136 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4" (UID: "e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:54:34 crc kubenswrapper[4787]: I1001 09:54:34.632019 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4-config-data" (OuterVolumeSpecName: "config-data") pod "e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4" (UID: "e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:54:34 crc kubenswrapper[4787]: I1001 09:54:34.692163 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:34 crc kubenswrapper[4787]: I1001 09:54:34.692460 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:34 crc kubenswrapper[4787]: I1001 09:54:34.692530 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-knchj\" (UniqueName: \"kubernetes.io/projected/e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4-kube-api-access-knchj\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:34 crc kubenswrapper[4787]: I1001 09:54:34.888919 4787 generic.go:334] "Generic (PLEG): container finished" podID="e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4" containerID="ab4386eec86d639ba0ad81456e5b1c2e571489777256838e3826cf0c99140712" exitCode=0 Oct 01 09:54:34 crc kubenswrapper[4787]: I1001 09:54:34.888983 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4","Type":"ContainerDied","Data":"ab4386eec86d639ba0ad81456e5b1c2e571489777256838e3826cf0c99140712"} Oct 01 09:54:34 crc kubenswrapper[4787]: I1001 09:54:34.889012 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4","Type":"ContainerDied","Data":"03eae54e5348fb9b7eb50caa9a87b7f97d89d1dbab696ea5117dbb0f5f73f7ce"} Oct 01 09:54:34 crc kubenswrapper[4787]: I1001 09:54:34.889029 4787 scope.go:117] "RemoveContainer" containerID="ab4386eec86d639ba0ad81456e5b1c2e571489777256838e3826cf0c99140712" Oct 01 09:54:34 crc kubenswrapper[4787]: I1001 09:54:34.889167 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 09:54:34 crc kubenswrapper[4787]: I1001 09:54:34.896611 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"06718a12-c4cb-43ed-be5f-0bee09272996","Type":"ContainerDied","Data":"b8527834f98426a9d4182b1b89990c2c65334ca925a0ce00e7503df61992661f"} Oct 01 09:54:34 crc kubenswrapper[4787]: I1001 09:54:34.896795 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 09:54:34 crc kubenswrapper[4787]: I1001 09:54:34.931267 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 01 09:54:34 crc kubenswrapper[4787]: I1001 09:54:34.942158 4787 scope.go:117] "RemoveContainer" containerID="ab4386eec86d639ba0ad81456e5b1c2e571489777256838e3826cf0c99140712" Oct 01 09:54:34 crc kubenswrapper[4787]: E1001 09:54:34.944974 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab4386eec86d639ba0ad81456e5b1c2e571489777256838e3826cf0c99140712\": container with ID starting with ab4386eec86d639ba0ad81456e5b1c2e571489777256838e3826cf0c99140712 not found: ID does not exist" containerID="ab4386eec86d639ba0ad81456e5b1c2e571489777256838e3826cf0c99140712" Oct 01 09:54:34 crc kubenswrapper[4787]: I1001 09:54:34.945014 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab4386eec86d639ba0ad81456e5b1c2e571489777256838e3826cf0c99140712"} err="failed to get container status \"ab4386eec86d639ba0ad81456e5b1c2e571489777256838e3826cf0c99140712\": rpc error: code = NotFound desc = could not find container \"ab4386eec86d639ba0ad81456e5b1c2e571489777256838e3826cf0c99140712\": container with ID starting with ab4386eec86d639ba0ad81456e5b1c2e571489777256838e3826cf0c99140712 not found: ID does not exist" Oct 01 09:54:34 crc kubenswrapper[4787]: I1001 09:54:34.945044 4787 scope.go:117] "RemoveContainer" containerID="72766222a72bbb3bc9a7b0d2308ee4218d433bed36ef7f8e8d0817cacabea125" Oct 01 09:54:34 crc kubenswrapper[4787]: I1001 09:54:34.952117 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 01 09:54:34 crc kubenswrapper[4787]: I1001 09:54:34.984453 4787 scope.go:117] "RemoveContainer" containerID="631ecef0955d0e36d72f5e7b8ed2bd3e36b84ae18a10e48d9c076aa7c2987b74" Oct 01 09:54:34 crc kubenswrapper[4787]: I1001 09:54:34.994237 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.018251 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.046446 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 01 09:54:35 crc kubenswrapper[4787]: E1001 09:54:35.046980 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4" containerName="nova-scheduler-scheduler" Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.047002 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4" containerName="nova-scheduler-scheduler" Oct 01 09:54:35 crc kubenswrapper[4787]: E1001 09:54:35.047040 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06718a12-c4cb-43ed-be5f-0bee09272996" containerName="nova-api-api" Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.047047 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="06718a12-c4cb-43ed-be5f-0bee09272996" containerName="nova-api-api" Oct 01 09:54:35 crc kubenswrapper[4787]: E1001 09:54:35.047092 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06718a12-c4cb-43ed-be5f-0bee09272996" containerName="nova-api-log" Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.047099 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="06718a12-c4cb-43ed-be5f-0bee09272996" containerName="nova-api-log" Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.047319 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4" containerName="nova-scheduler-scheduler" Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.047338 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="06718a12-c4cb-43ed-be5f-0bee09272996" containerName="nova-api-api" Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.047354 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="06718a12-c4cb-43ed-be5f-0bee09272996" containerName="nova-api-log" Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.049488 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.050616 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.050786 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.054211 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.056610 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.056837 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.074955 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.103297 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec425a80-032d-48b0-b46d-a38d98482149-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ec425a80-032d-48b0-b46d-a38d98482149\") " pod="openstack/nova-api-0" Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.103352 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec425a80-032d-48b0-b46d-a38d98482149-logs\") pod \"nova-api-0\" (UID: \"ec425a80-032d-48b0-b46d-a38d98482149\") " pod="openstack/nova-api-0" Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.103407 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c15f26a3-6250-43c9-822a-974a2f3c0a63-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c15f26a3-6250-43c9-822a-974a2f3c0a63\") " pod="openstack/nova-scheduler-0" Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.103449 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c15f26a3-6250-43c9-822a-974a2f3c0a63-config-data\") pod \"nova-scheduler-0\" (UID: \"c15f26a3-6250-43c9-822a-974a2f3c0a63\") " pod="openstack/nova-scheduler-0" Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.103474 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdfbk\" (UniqueName: \"kubernetes.io/projected/c15f26a3-6250-43c9-822a-974a2f3c0a63-kube-api-access-qdfbk\") pod \"nova-scheduler-0\" (UID: \"c15f26a3-6250-43c9-822a-974a2f3c0a63\") " pod="openstack/nova-scheduler-0" Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.103524 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77hcf\" (UniqueName: \"kubernetes.io/projected/ec425a80-032d-48b0-b46d-a38d98482149-kube-api-access-77hcf\") pod \"nova-api-0\" (UID: \"ec425a80-032d-48b0-b46d-a38d98482149\") " pod="openstack/nova-api-0" Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.103541 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec425a80-032d-48b0-b46d-a38d98482149-config-data\") pod \"nova-api-0\" (UID: \"ec425a80-032d-48b0-b46d-a38d98482149\") " pod="openstack/nova-api-0" Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.206208 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec425a80-032d-48b0-b46d-a38d98482149-logs\") pod \"nova-api-0\" (UID: \"ec425a80-032d-48b0-b46d-a38d98482149\") " pod="openstack/nova-api-0" Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.206297 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c15f26a3-6250-43c9-822a-974a2f3c0a63-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c15f26a3-6250-43c9-822a-974a2f3c0a63\") " pod="openstack/nova-scheduler-0" Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.206345 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c15f26a3-6250-43c9-822a-974a2f3c0a63-config-data\") pod \"nova-scheduler-0\" (UID: \"c15f26a3-6250-43c9-822a-974a2f3c0a63\") " pod="openstack/nova-scheduler-0" Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.206375 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdfbk\" (UniqueName: \"kubernetes.io/projected/c15f26a3-6250-43c9-822a-974a2f3c0a63-kube-api-access-qdfbk\") pod \"nova-scheduler-0\" (UID: \"c15f26a3-6250-43c9-822a-974a2f3c0a63\") " pod="openstack/nova-scheduler-0" Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.206429 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77hcf\" (UniqueName: \"kubernetes.io/projected/ec425a80-032d-48b0-b46d-a38d98482149-kube-api-access-77hcf\") pod \"nova-api-0\" (UID: \"ec425a80-032d-48b0-b46d-a38d98482149\") " pod="openstack/nova-api-0" Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.206449 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec425a80-032d-48b0-b46d-a38d98482149-config-data\") pod \"nova-api-0\" (UID: \"ec425a80-032d-48b0-b46d-a38d98482149\") " pod="openstack/nova-api-0" Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.206497 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec425a80-032d-48b0-b46d-a38d98482149-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ec425a80-032d-48b0-b46d-a38d98482149\") " pod="openstack/nova-api-0" Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.210542 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec425a80-032d-48b0-b46d-a38d98482149-logs\") pod \"nova-api-0\" (UID: \"ec425a80-032d-48b0-b46d-a38d98482149\") " pod="openstack/nova-api-0" Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.224997 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec425a80-032d-48b0-b46d-a38d98482149-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ec425a80-032d-48b0-b46d-a38d98482149\") " pod="openstack/nova-api-0" Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.226854 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec425a80-032d-48b0-b46d-a38d98482149-config-data\") pod \"nova-api-0\" (UID: \"ec425a80-032d-48b0-b46d-a38d98482149\") " pod="openstack/nova-api-0" Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.229427 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c15f26a3-6250-43c9-822a-974a2f3c0a63-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c15f26a3-6250-43c9-822a-974a2f3c0a63\") " pod="openstack/nova-scheduler-0" Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.230686 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c15f26a3-6250-43c9-822a-974a2f3c0a63-config-data\") pod \"nova-scheduler-0\" (UID: \"c15f26a3-6250-43c9-822a-974a2f3c0a63\") " pod="openstack/nova-scheduler-0" Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.232297 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77hcf\" (UniqueName: \"kubernetes.io/projected/ec425a80-032d-48b0-b46d-a38d98482149-kube-api-access-77hcf\") pod \"nova-api-0\" (UID: \"ec425a80-032d-48b0-b46d-a38d98482149\") " pod="openstack/nova-api-0" Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.232987 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdfbk\" (UniqueName: \"kubernetes.io/projected/c15f26a3-6250-43c9-822a-974a2f3c0a63-kube-api-access-qdfbk\") pod \"nova-scheduler-0\" (UID: \"c15f26a3-6250-43c9-822a-974a2f3c0a63\") " pod="openstack/nova-scheduler-0" Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.380179 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.397773 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.878482 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.920496 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c15f26a3-6250-43c9-822a-974a2f3c0a63","Type":"ContainerStarted","Data":"ea86e3d2a7445df104cd9e94669cebbdf1db437a974770aabfcde27ec96380bc"} Oct 01 09:54:35 crc kubenswrapper[4787]: I1001 09:54:35.954072 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 09:54:36 crc kubenswrapper[4787]: I1001 09:54:36.537402 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06718a12-c4cb-43ed-be5f-0bee09272996" path="/var/lib/kubelet/pods/06718a12-c4cb-43ed-be5f-0bee09272996/volumes" Oct 01 09:54:36 crc kubenswrapper[4787]: I1001 09:54:36.538511 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4" path="/var/lib/kubelet/pods/e2372a84-e50b-4b96-a8e7-e4a0bfb9e1b4/volumes" Oct 01 09:54:36 crc kubenswrapper[4787]: I1001 09:54:36.937191 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c15f26a3-6250-43c9-822a-974a2f3c0a63","Type":"ContainerStarted","Data":"e1033e70a9ecb6214b67b36775d24b8575db35d5df073e0b8ff0126db0442282"} Oct 01 09:54:36 crc kubenswrapper[4787]: I1001 09:54:36.945928 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ec425a80-032d-48b0-b46d-a38d98482149","Type":"ContainerStarted","Data":"47877f416671132ec65ad4daf6c9681db228d733d74960a9216ae06960eed683"} Oct 01 09:54:36 crc kubenswrapper[4787]: I1001 09:54:36.945983 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ec425a80-032d-48b0-b46d-a38d98482149","Type":"ContainerStarted","Data":"aac3eccdc8574c073088fc9f79a2eb36d2a4b971b79fead4de1bf5723a6378a6"} Oct 01 09:54:36 crc kubenswrapper[4787]: I1001 09:54:36.945995 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ec425a80-032d-48b0-b46d-a38d98482149","Type":"ContainerStarted","Data":"57b2d5c54eff49731d44a6c954a5c971dbb1ac37debcf258b1a81ab5847b4101"} Oct 01 09:54:36 crc kubenswrapper[4787]: I1001 09:54:36.962172 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.962150962 podStartE2EDuration="2.962150962s" podCreationTimestamp="2025-10-01 09:54:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:54:36.961069564 +0000 UTC m=+1109.076213731" watchObservedRunningTime="2025-10-01 09:54:36.962150962 +0000 UTC m=+1109.077295139" Oct 01 09:54:36 crc kubenswrapper[4787]: I1001 09:54:36.979648 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.979631565 podStartE2EDuration="2.979631565s" podCreationTimestamp="2025-10-01 09:54:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:54:36.978967299 +0000 UTC m=+1109.094111456" watchObservedRunningTime="2025-10-01 09:54:36.979631565 +0000 UTC m=+1109.094775722" Oct 01 09:54:37 crc kubenswrapper[4787]: I1001 09:54:37.552905 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 09:54:37 crc kubenswrapper[4787]: I1001 09:54:37.553548 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="469f4dc6-d176-4497-ac69-6130a375737b" containerName="kube-state-metrics" containerID="cri-o://e88e716b091e96dfbeec718541083c5b8f296f8f98fea8a472a2917ee8c45f46" gracePeriod=30 Oct 01 09:54:37 crc kubenswrapper[4787]: I1001 09:54:37.958510 4787 generic.go:334] "Generic (PLEG): container finished" podID="469f4dc6-d176-4497-ac69-6130a375737b" containerID="e88e716b091e96dfbeec718541083c5b8f296f8f98fea8a472a2917ee8c45f46" exitCode=2 Oct 01 09:54:37 crc kubenswrapper[4787]: I1001 09:54:37.958553 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"469f4dc6-d176-4497-ac69-6130a375737b","Type":"ContainerDied","Data":"e88e716b091e96dfbeec718541083c5b8f296f8f98fea8a472a2917ee8c45f46"} Oct 01 09:54:37 crc kubenswrapper[4787]: I1001 09:54:37.958872 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"469f4dc6-d176-4497-ac69-6130a375737b","Type":"ContainerDied","Data":"1f63527c8f398eb15518a149dbb2c1123e7fe6b54914e56a50fa3f06f07cb2fa"} Oct 01 09:54:37 crc kubenswrapper[4787]: I1001 09:54:37.958886 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f63527c8f398eb15518a149dbb2c1123e7fe6b54914e56a50fa3f06f07cb2fa" Oct 01 09:54:38 crc kubenswrapper[4787]: I1001 09:54:38.030021 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 01 09:54:38 crc kubenswrapper[4787]: I1001 09:54:38.072430 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wj8f6\" (UniqueName: \"kubernetes.io/projected/469f4dc6-d176-4497-ac69-6130a375737b-kube-api-access-wj8f6\") pod \"469f4dc6-d176-4497-ac69-6130a375737b\" (UID: \"469f4dc6-d176-4497-ac69-6130a375737b\") " Oct 01 09:54:38 crc kubenswrapper[4787]: I1001 09:54:38.080418 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/469f4dc6-d176-4497-ac69-6130a375737b-kube-api-access-wj8f6" (OuterVolumeSpecName: "kube-api-access-wj8f6") pod "469f4dc6-d176-4497-ac69-6130a375737b" (UID: "469f4dc6-d176-4497-ac69-6130a375737b"). InnerVolumeSpecName "kube-api-access-wj8f6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:54:38 crc kubenswrapper[4787]: I1001 09:54:38.174229 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wj8f6\" (UniqueName: \"kubernetes.io/projected/469f4dc6-d176-4497-ac69-6130a375737b-kube-api-access-wj8f6\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:38 crc kubenswrapper[4787]: I1001 09:54:38.969366 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 01 09:54:39 crc kubenswrapper[4787]: I1001 09:54:39.001392 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 09:54:39 crc kubenswrapper[4787]: I1001 09:54:39.018289 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 09:54:39 crc kubenswrapper[4787]: I1001 09:54:39.033176 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 09:54:39 crc kubenswrapper[4787]: E1001 09:54:39.033722 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="469f4dc6-d176-4497-ac69-6130a375737b" containerName="kube-state-metrics" Oct 01 09:54:39 crc kubenswrapper[4787]: I1001 09:54:39.033744 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="469f4dc6-d176-4497-ac69-6130a375737b" containerName="kube-state-metrics" Oct 01 09:54:39 crc kubenswrapper[4787]: I1001 09:54:39.033914 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="469f4dc6-d176-4497-ac69-6130a375737b" containerName="kube-state-metrics" Oct 01 09:54:39 crc kubenswrapper[4787]: I1001 09:54:39.034600 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 01 09:54:39 crc kubenswrapper[4787]: I1001 09:54:39.037139 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 01 09:54:39 crc kubenswrapper[4787]: I1001 09:54:39.038011 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 01 09:54:39 crc kubenswrapper[4787]: I1001 09:54:39.041457 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 09:54:39 crc kubenswrapper[4787]: I1001 09:54:39.100869 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1c9f3e0-fa6b-4ef4-ae4a-a0aea3bfc262-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"c1c9f3e0-fa6b-4ef4-ae4a-a0aea3bfc262\") " pod="openstack/kube-state-metrics-0" Oct 01 09:54:39 crc kubenswrapper[4787]: I1001 09:54:39.101264 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/c1c9f3e0-fa6b-4ef4-ae4a-a0aea3bfc262-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"c1c9f3e0-fa6b-4ef4-ae4a-a0aea3bfc262\") " pod="openstack/kube-state-metrics-0" Oct 01 09:54:39 crc kubenswrapper[4787]: I1001 09:54:39.101475 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jh2x\" (UniqueName: \"kubernetes.io/projected/c1c9f3e0-fa6b-4ef4-ae4a-a0aea3bfc262-kube-api-access-7jh2x\") pod \"kube-state-metrics-0\" (UID: \"c1c9f3e0-fa6b-4ef4-ae4a-a0aea3bfc262\") " pod="openstack/kube-state-metrics-0" Oct 01 09:54:39 crc kubenswrapper[4787]: I1001 09:54:39.101646 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1c9f3e0-fa6b-4ef4-ae4a-a0aea3bfc262-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"c1c9f3e0-fa6b-4ef4-ae4a-a0aea3bfc262\") " pod="openstack/kube-state-metrics-0" Oct 01 09:54:39 crc kubenswrapper[4787]: I1001 09:54:39.203589 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/c1c9f3e0-fa6b-4ef4-ae4a-a0aea3bfc262-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"c1c9f3e0-fa6b-4ef4-ae4a-a0aea3bfc262\") " pod="openstack/kube-state-metrics-0" Oct 01 09:54:39 crc kubenswrapper[4787]: I1001 09:54:39.203705 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jh2x\" (UniqueName: \"kubernetes.io/projected/c1c9f3e0-fa6b-4ef4-ae4a-a0aea3bfc262-kube-api-access-7jh2x\") pod \"kube-state-metrics-0\" (UID: \"c1c9f3e0-fa6b-4ef4-ae4a-a0aea3bfc262\") " pod="openstack/kube-state-metrics-0" Oct 01 09:54:39 crc kubenswrapper[4787]: I1001 09:54:39.203774 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1c9f3e0-fa6b-4ef4-ae4a-a0aea3bfc262-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"c1c9f3e0-fa6b-4ef4-ae4a-a0aea3bfc262\") " pod="openstack/kube-state-metrics-0" Oct 01 09:54:39 crc kubenswrapper[4787]: I1001 09:54:39.203837 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1c9f3e0-fa6b-4ef4-ae4a-a0aea3bfc262-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"c1c9f3e0-fa6b-4ef4-ae4a-a0aea3bfc262\") " pod="openstack/kube-state-metrics-0" Oct 01 09:54:39 crc kubenswrapper[4787]: I1001 09:54:39.207852 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1c9f3e0-fa6b-4ef4-ae4a-a0aea3bfc262-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"c1c9f3e0-fa6b-4ef4-ae4a-a0aea3bfc262\") " pod="openstack/kube-state-metrics-0" Oct 01 09:54:39 crc kubenswrapper[4787]: I1001 09:54:39.208056 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1c9f3e0-fa6b-4ef4-ae4a-a0aea3bfc262-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"c1c9f3e0-fa6b-4ef4-ae4a-a0aea3bfc262\") " pod="openstack/kube-state-metrics-0" Oct 01 09:54:39 crc kubenswrapper[4787]: I1001 09:54:39.223415 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jh2x\" (UniqueName: \"kubernetes.io/projected/c1c9f3e0-fa6b-4ef4-ae4a-a0aea3bfc262-kube-api-access-7jh2x\") pod \"kube-state-metrics-0\" (UID: \"c1c9f3e0-fa6b-4ef4-ae4a-a0aea3bfc262\") " pod="openstack/kube-state-metrics-0" Oct 01 09:54:39 crc kubenswrapper[4787]: I1001 09:54:39.224621 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/c1c9f3e0-fa6b-4ef4-ae4a-a0aea3bfc262-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"c1c9f3e0-fa6b-4ef4-ae4a-a0aea3bfc262\") " pod="openstack/kube-state-metrics-0" Oct 01 09:54:39 crc kubenswrapper[4787]: I1001 09:54:39.355520 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 01 09:54:39 crc kubenswrapper[4787]: I1001 09:54:39.439236 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:54:39 crc kubenswrapper[4787]: I1001 09:54:39.439550 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5c67fad8-4bad-49ac-bd9f-218c6d0d98d5" containerName="ceilometer-central-agent" containerID="cri-o://e338c2385f03560491eccb8129f3be389c57a3b6d485da5c158a9038e5530863" gracePeriod=30 Oct 01 09:54:39 crc kubenswrapper[4787]: I1001 09:54:39.439711 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5c67fad8-4bad-49ac-bd9f-218c6d0d98d5" containerName="proxy-httpd" containerID="cri-o://ada12ec4a7e7f6e31663ea7ee1e674967e3d7259c5c0ea8efc31cc2b343291e1" gracePeriod=30 Oct 01 09:54:39 crc kubenswrapper[4787]: I1001 09:54:39.439752 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5c67fad8-4bad-49ac-bd9f-218c6d0d98d5" containerName="sg-core" containerID="cri-o://d1a6aa02188231959ecffe4fb8f1113d342324f03ac38e6084c61225e4e7579c" gracePeriod=30 Oct 01 09:54:39 crc kubenswrapper[4787]: I1001 09:54:39.439784 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5c67fad8-4bad-49ac-bd9f-218c6d0d98d5" containerName="ceilometer-notification-agent" containerID="cri-o://5ba6561fb8134d0a04c50d0e599b609eab84e3607f9c90d31673b339326eda2f" gracePeriod=30 Oct 01 09:54:39 crc kubenswrapper[4787]: I1001 09:54:39.840432 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 01 09:54:39 crc kubenswrapper[4787]: W1001 09:54:39.848922 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc1c9f3e0_fa6b_4ef4_ae4a_a0aea3bfc262.slice/crio-58bd9375593ac71c1d3d4b9b11fe2a86c74e10dda558ba8300519feeb2d99527 WatchSource:0}: Error finding container 58bd9375593ac71c1d3d4b9b11fe2a86c74e10dda558ba8300519feeb2d99527: Status 404 returned error can't find the container with id 58bd9375593ac71c1d3d4b9b11fe2a86c74e10dda558ba8300519feeb2d99527 Oct 01 09:54:39 crc kubenswrapper[4787]: I1001 09:54:39.981162 4787 generic.go:334] "Generic (PLEG): container finished" podID="5c67fad8-4bad-49ac-bd9f-218c6d0d98d5" containerID="ada12ec4a7e7f6e31663ea7ee1e674967e3d7259c5c0ea8efc31cc2b343291e1" exitCode=0 Oct 01 09:54:39 crc kubenswrapper[4787]: I1001 09:54:39.981202 4787 generic.go:334] "Generic (PLEG): container finished" podID="5c67fad8-4bad-49ac-bd9f-218c6d0d98d5" containerID="d1a6aa02188231959ecffe4fb8f1113d342324f03ac38e6084c61225e4e7579c" exitCode=2 Oct 01 09:54:39 crc kubenswrapper[4787]: I1001 09:54:39.981215 4787 generic.go:334] "Generic (PLEG): container finished" podID="5c67fad8-4bad-49ac-bd9f-218c6d0d98d5" containerID="e338c2385f03560491eccb8129f3be389c57a3b6d485da5c158a9038e5530863" exitCode=0 Oct 01 09:54:39 crc kubenswrapper[4787]: I1001 09:54:39.981262 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5","Type":"ContainerDied","Data":"ada12ec4a7e7f6e31663ea7ee1e674967e3d7259c5c0ea8efc31cc2b343291e1"} Oct 01 09:54:39 crc kubenswrapper[4787]: I1001 09:54:39.981296 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5","Type":"ContainerDied","Data":"d1a6aa02188231959ecffe4fb8f1113d342324f03ac38e6084c61225e4e7579c"} Oct 01 09:54:39 crc kubenswrapper[4787]: I1001 09:54:39.981309 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5","Type":"ContainerDied","Data":"e338c2385f03560491eccb8129f3be389c57a3b6d485da5c158a9038e5530863"} Oct 01 09:54:39 crc kubenswrapper[4787]: I1001 09:54:39.982355 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c1c9f3e0-fa6b-4ef4-ae4a-a0aea3bfc262","Type":"ContainerStarted","Data":"58bd9375593ac71c1d3d4b9b11fe2a86c74e10dda558ba8300519feeb2d99527"} Oct 01 09:54:40 crc kubenswrapper[4787]: I1001 09:54:40.266931 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 01 09:54:40 crc kubenswrapper[4787]: I1001 09:54:40.381293 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 01 09:54:40 crc kubenswrapper[4787]: I1001 09:54:40.537177 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="469f4dc6-d176-4497-ac69-6130a375737b" path="/var/lib/kubelet/pods/469f4dc6-d176-4497-ac69-6130a375737b/volumes" Oct 01 09:54:40 crc kubenswrapper[4787]: I1001 09:54:40.992916 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c1c9f3e0-fa6b-4ef4-ae4a-a0aea3bfc262","Type":"ContainerStarted","Data":"1a19acccbae3e9525190b2a942f72bf3f624f397008666b847dc4f756ffaa0c0"} Oct 01 09:54:40 crc kubenswrapper[4787]: I1001 09:54:40.993438 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 01 09:54:41 crc kubenswrapper[4787]: I1001 09:54:41.028157 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.675097215 podStartE2EDuration="3.028135435s" podCreationTimestamp="2025-10-01 09:54:38 +0000 UTC" firstStartedPulling="2025-10-01 09:54:39.864891832 +0000 UTC m=+1111.980035989" lastFinishedPulling="2025-10-01 09:54:40.217930052 +0000 UTC m=+1112.333074209" observedRunningTime="2025-10-01 09:54:41.006380663 +0000 UTC m=+1113.121524830" watchObservedRunningTime="2025-10-01 09:54:41.028135435 +0000 UTC m=+1113.143279592" Oct 01 09:54:41 crc kubenswrapper[4787]: I1001 09:54:41.250936 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:54:41 crc kubenswrapper[4787]: I1001 09:54:41.251037 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:54:41 crc kubenswrapper[4787]: I1001 09:54:41.251138 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" Oct 01 09:54:41 crc kubenswrapper[4787]: I1001 09:54:41.252228 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"86d3a42fa94bb1529366685be8d4c187b77d63f0ed73a16bbe3812ceaa99b7ae"} pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 09:54:41 crc kubenswrapper[4787]: I1001 09:54:41.252337 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" containerID="cri-o://86d3a42fa94bb1529366685be8d4c187b77d63f0ed73a16bbe3812ceaa99b7ae" gracePeriod=600 Oct 01 09:54:42 crc kubenswrapper[4787]: I1001 09:54:42.009253 4787 generic.go:334] "Generic (PLEG): container finished" podID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerID="86d3a42fa94bb1529366685be8d4c187b77d63f0ed73a16bbe3812ceaa99b7ae" exitCode=0 Oct 01 09:54:42 crc kubenswrapper[4787]: I1001 09:54:42.009331 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" event={"ID":"fa51b883-5f3e-4141-9d4c-37704aac7718","Type":"ContainerDied","Data":"86d3a42fa94bb1529366685be8d4c187b77d63f0ed73a16bbe3812ceaa99b7ae"} Oct 01 09:54:42 crc kubenswrapper[4787]: I1001 09:54:42.009973 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" event={"ID":"fa51b883-5f3e-4141-9d4c-37704aac7718","Type":"ContainerStarted","Data":"a85043035562ba4be7fcc084d6a9297752334c5b023237849bd5d1037f591d75"} Oct 01 09:54:42 crc kubenswrapper[4787]: I1001 09:54:42.010019 4787 scope.go:117] "RemoveContainer" containerID="d423136db78a796507e81b181b34f0022442acd501d56cba0addad4da7455409" Oct 01 09:54:42 crc kubenswrapper[4787]: I1001 09:54:42.767657 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:54:42 crc kubenswrapper[4787]: I1001 09:54:42.873303 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-config-data\") pod \"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5\" (UID: \"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5\") " Oct 01 09:54:42 crc kubenswrapper[4787]: I1001 09:54:42.873377 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-log-httpd\") pod \"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5\" (UID: \"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5\") " Oct 01 09:54:42 crc kubenswrapper[4787]: I1001 09:54:42.873511 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-sg-core-conf-yaml\") pod \"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5\" (UID: \"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5\") " Oct 01 09:54:42 crc kubenswrapper[4787]: I1001 09:54:42.873529 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-combined-ca-bundle\") pod \"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5\" (UID: \"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5\") " Oct 01 09:54:42 crc kubenswrapper[4787]: I1001 09:54:42.873582 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grwtt\" (UniqueName: \"kubernetes.io/projected/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-kube-api-access-grwtt\") pod \"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5\" (UID: \"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5\") " Oct 01 09:54:42 crc kubenswrapper[4787]: I1001 09:54:42.873612 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-scripts\") pod \"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5\" (UID: \"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5\") " Oct 01 09:54:42 crc kubenswrapper[4787]: I1001 09:54:42.873659 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-run-httpd\") pod \"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5\" (UID: \"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5\") " Oct 01 09:54:42 crc kubenswrapper[4787]: I1001 09:54:42.874120 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5c67fad8-4bad-49ac-bd9f-218c6d0d98d5" (UID: "5c67fad8-4bad-49ac-bd9f-218c6d0d98d5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:54:42 crc kubenswrapper[4787]: I1001 09:54:42.874602 4787 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:42 crc kubenswrapper[4787]: I1001 09:54:42.879445 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5c67fad8-4bad-49ac-bd9f-218c6d0d98d5" (UID: "5c67fad8-4bad-49ac-bd9f-218c6d0d98d5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:54:42 crc kubenswrapper[4787]: I1001 09:54:42.881363 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-kube-api-access-grwtt" (OuterVolumeSpecName: "kube-api-access-grwtt") pod "5c67fad8-4bad-49ac-bd9f-218c6d0d98d5" (UID: "5c67fad8-4bad-49ac-bd9f-218c6d0d98d5"). InnerVolumeSpecName "kube-api-access-grwtt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:54:42 crc kubenswrapper[4787]: I1001 09:54:42.892950 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-scripts" (OuterVolumeSpecName: "scripts") pod "5c67fad8-4bad-49ac-bd9f-218c6d0d98d5" (UID: "5c67fad8-4bad-49ac-bd9f-218c6d0d98d5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:54:42 crc kubenswrapper[4787]: I1001 09:54:42.904423 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5c67fad8-4bad-49ac-bd9f-218c6d0d98d5" (UID: "5c67fad8-4bad-49ac-bd9f-218c6d0d98d5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:54:42 crc kubenswrapper[4787]: I1001 09:54:42.961296 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5c67fad8-4bad-49ac-bd9f-218c6d0d98d5" (UID: "5c67fad8-4bad-49ac-bd9f-218c6d0d98d5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:54:42 crc kubenswrapper[4787]: I1001 09:54:42.976131 4787 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:42 crc kubenswrapper[4787]: I1001 09:54:42.976161 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:42 crc kubenswrapper[4787]: I1001 09:54:42.976171 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grwtt\" (UniqueName: \"kubernetes.io/projected/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-kube-api-access-grwtt\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:42 crc kubenswrapper[4787]: I1001 09:54:42.976182 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:42 crc kubenswrapper[4787]: I1001 09:54:42.976190 4787 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:42 crc kubenswrapper[4787]: I1001 09:54:42.996995 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-config-data" (OuterVolumeSpecName: "config-data") pod "5c67fad8-4bad-49ac-bd9f-218c6d0d98d5" (UID: "5c67fad8-4bad-49ac-bd9f-218c6d0d98d5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.024015 4787 generic.go:334] "Generic (PLEG): container finished" podID="5c67fad8-4bad-49ac-bd9f-218c6d0d98d5" containerID="5ba6561fb8134d0a04c50d0e599b609eab84e3607f9c90d31673b339326eda2f" exitCode=0 Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.024100 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5","Type":"ContainerDied","Data":"5ba6561fb8134d0a04c50d0e599b609eab84e3607f9c90d31673b339326eda2f"} Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.024138 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5c67fad8-4bad-49ac-bd9f-218c6d0d98d5","Type":"ContainerDied","Data":"b3e27dd3d23025dfe522f4d751d860646c014af5bc392ca7241cd4bdf30c3149"} Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.024165 4787 scope.go:117] "RemoveContainer" containerID="ada12ec4a7e7f6e31663ea7ee1e674967e3d7259c5c0ea8efc31cc2b343291e1" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.024172 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.048695 4787 scope.go:117] "RemoveContainer" containerID="d1a6aa02188231959ecffe4fb8f1113d342324f03ac38e6084c61225e4e7579c" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.066485 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.078427 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.082108 4787 scope.go:117] "RemoveContainer" containerID="5ba6561fb8134d0a04c50d0e599b609eab84e3607f9c90d31673b339326eda2f" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.087598 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.095570 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:54:43 crc kubenswrapper[4787]: E1001 09:54:43.096041 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c67fad8-4bad-49ac-bd9f-218c6d0d98d5" containerName="proxy-httpd" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.096058 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c67fad8-4bad-49ac-bd9f-218c6d0d98d5" containerName="proxy-httpd" Oct 01 09:54:43 crc kubenswrapper[4787]: E1001 09:54:43.096070 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c67fad8-4bad-49ac-bd9f-218c6d0d98d5" containerName="ceilometer-notification-agent" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.096087 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c67fad8-4bad-49ac-bd9f-218c6d0d98d5" containerName="ceilometer-notification-agent" Oct 01 09:54:43 crc kubenswrapper[4787]: E1001 09:54:43.096104 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c67fad8-4bad-49ac-bd9f-218c6d0d98d5" containerName="sg-core" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.096110 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c67fad8-4bad-49ac-bd9f-218c6d0d98d5" containerName="sg-core" Oct 01 09:54:43 crc kubenswrapper[4787]: E1001 09:54:43.096140 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c67fad8-4bad-49ac-bd9f-218c6d0d98d5" containerName="ceilometer-central-agent" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.096146 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c67fad8-4bad-49ac-bd9f-218c6d0d98d5" containerName="ceilometer-central-agent" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.096341 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c67fad8-4bad-49ac-bd9f-218c6d0d98d5" containerName="ceilometer-central-agent" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.096356 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c67fad8-4bad-49ac-bd9f-218c6d0d98d5" containerName="proxy-httpd" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.096368 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c67fad8-4bad-49ac-bd9f-218c6d0d98d5" containerName="sg-core" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.096387 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c67fad8-4bad-49ac-bd9f-218c6d0d98d5" containerName="ceilometer-notification-agent" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.098277 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.099977 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.100234 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.100459 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.103900 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.130700 4787 scope.go:117] "RemoveContainer" containerID="e338c2385f03560491eccb8129f3be389c57a3b6d485da5c158a9038e5530863" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.151411 4787 scope.go:117] "RemoveContainer" containerID="ada12ec4a7e7f6e31663ea7ee1e674967e3d7259c5c0ea8efc31cc2b343291e1" Oct 01 09:54:43 crc kubenswrapper[4787]: E1001 09:54:43.151909 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ada12ec4a7e7f6e31663ea7ee1e674967e3d7259c5c0ea8efc31cc2b343291e1\": container with ID starting with ada12ec4a7e7f6e31663ea7ee1e674967e3d7259c5c0ea8efc31cc2b343291e1 not found: ID does not exist" containerID="ada12ec4a7e7f6e31663ea7ee1e674967e3d7259c5c0ea8efc31cc2b343291e1" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.152021 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ada12ec4a7e7f6e31663ea7ee1e674967e3d7259c5c0ea8efc31cc2b343291e1"} err="failed to get container status \"ada12ec4a7e7f6e31663ea7ee1e674967e3d7259c5c0ea8efc31cc2b343291e1\": rpc error: code = NotFound desc = could not find container \"ada12ec4a7e7f6e31663ea7ee1e674967e3d7259c5c0ea8efc31cc2b343291e1\": container with ID starting with ada12ec4a7e7f6e31663ea7ee1e674967e3d7259c5c0ea8efc31cc2b343291e1 not found: ID does not exist" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.152103 4787 scope.go:117] "RemoveContainer" containerID="d1a6aa02188231959ecffe4fb8f1113d342324f03ac38e6084c61225e4e7579c" Oct 01 09:54:43 crc kubenswrapper[4787]: E1001 09:54:43.152554 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1a6aa02188231959ecffe4fb8f1113d342324f03ac38e6084c61225e4e7579c\": container with ID starting with d1a6aa02188231959ecffe4fb8f1113d342324f03ac38e6084c61225e4e7579c not found: ID does not exist" containerID="d1a6aa02188231959ecffe4fb8f1113d342324f03ac38e6084c61225e4e7579c" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.152603 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1a6aa02188231959ecffe4fb8f1113d342324f03ac38e6084c61225e4e7579c"} err="failed to get container status \"d1a6aa02188231959ecffe4fb8f1113d342324f03ac38e6084c61225e4e7579c\": rpc error: code = NotFound desc = could not find container \"d1a6aa02188231959ecffe4fb8f1113d342324f03ac38e6084c61225e4e7579c\": container with ID starting with d1a6aa02188231959ecffe4fb8f1113d342324f03ac38e6084c61225e4e7579c not found: ID does not exist" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.152627 4787 scope.go:117] "RemoveContainer" containerID="5ba6561fb8134d0a04c50d0e599b609eab84e3607f9c90d31673b339326eda2f" Oct 01 09:54:43 crc kubenswrapper[4787]: E1001 09:54:43.152908 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ba6561fb8134d0a04c50d0e599b609eab84e3607f9c90d31673b339326eda2f\": container with ID starting with 5ba6561fb8134d0a04c50d0e599b609eab84e3607f9c90d31673b339326eda2f not found: ID does not exist" containerID="5ba6561fb8134d0a04c50d0e599b609eab84e3607f9c90d31673b339326eda2f" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.152936 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ba6561fb8134d0a04c50d0e599b609eab84e3607f9c90d31673b339326eda2f"} err="failed to get container status \"5ba6561fb8134d0a04c50d0e599b609eab84e3607f9c90d31673b339326eda2f\": rpc error: code = NotFound desc = could not find container \"5ba6561fb8134d0a04c50d0e599b609eab84e3607f9c90d31673b339326eda2f\": container with ID starting with 5ba6561fb8134d0a04c50d0e599b609eab84e3607f9c90d31673b339326eda2f not found: ID does not exist" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.152953 4787 scope.go:117] "RemoveContainer" containerID="e338c2385f03560491eccb8129f3be389c57a3b6d485da5c158a9038e5530863" Oct 01 09:54:43 crc kubenswrapper[4787]: E1001 09:54:43.153457 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e338c2385f03560491eccb8129f3be389c57a3b6d485da5c158a9038e5530863\": container with ID starting with e338c2385f03560491eccb8129f3be389c57a3b6d485da5c158a9038e5530863 not found: ID does not exist" containerID="e338c2385f03560491eccb8129f3be389c57a3b6d485da5c158a9038e5530863" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.153484 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e338c2385f03560491eccb8129f3be389c57a3b6d485da5c158a9038e5530863"} err="failed to get container status \"e338c2385f03560491eccb8129f3be389c57a3b6d485da5c158a9038e5530863\": rpc error: code = NotFound desc = could not find container \"e338c2385f03560491eccb8129f3be389c57a3b6d485da5c158a9038e5530863\": container with ID starting with e338c2385f03560491eccb8129f3be389c57a3b6d485da5c158a9038e5530863 not found: ID does not exist" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.281213 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87c34fde-d10d-4c46-a421-80e0ec259618-config-data\") pod \"ceilometer-0\" (UID: \"87c34fde-d10d-4c46-a421-80e0ec259618\") " pod="openstack/ceilometer-0" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.281529 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87c34fde-d10d-4c46-a421-80e0ec259618-scripts\") pod \"ceilometer-0\" (UID: \"87c34fde-d10d-4c46-a421-80e0ec259618\") " pod="openstack/ceilometer-0" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.281625 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q27rn\" (UniqueName: \"kubernetes.io/projected/87c34fde-d10d-4c46-a421-80e0ec259618-kube-api-access-q27rn\") pod \"ceilometer-0\" (UID: \"87c34fde-d10d-4c46-a421-80e0ec259618\") " pod="openstack/ceilometer-0" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.281729 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87c34fde-d10d-4c46-a421-80e0ec259618-run-httpd\") pod \"ceilometer-0\" (UID: \"87c34fde-d10d-4c46-a421-80e0ec259618\") " pod="openstack/ceilometer-0" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.281848 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87c34fde-d10d-4c46-a421-80e0ec259618-log-httpd\") pod \"ceilometer-0\" (UID: \"87c34fde-d10d-4c46-a421-80e0ec259618\") " pod="openstack/ceilometer-0" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.281925 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c34fde-d10d-4c46-a421-80e0ec259618-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"87c34fde-d10d-4c46-a421-80e0ec259618\") " pod="openstack/ceilometer-0" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.282011 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87c34fde-d10d-4c46-a421-80e0ec259618-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"87c34fde-d10d-4c46-a421-80e0ec259618\") " pod="openstack/ceilometer-0" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.282149 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/87c34fde-d10d-4c46-a421-80e0ec259618-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"87c34fde-d10d-4c46-a421-80e0ec259618\") " pod="openstack/ceilometer-0" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.385258 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87c34fde-d10d-4c46-a421-80e0ec259618-log-httpd\") pod \"ceilometer-0\" (UID: \"87c34fde-d10d-4c46-a421-80e0ec259618\") " pod="openstack/ceilometer-0" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.385308 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c34fde-d10d-4c46-a421-80e0ec259618-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"87c34fde-d10d-4c46-a421-80e0ec259618\") " pod="openstack/ceilometer-0" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.385356 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87c34fde-d10d-4c46-a421-80e0ec259618-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"87c34fde-d10d-4c46-a421-80e0ec259618\") " pod="openstack/ceilometer-0" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.385387 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/87c34fde-d10d-4c46-a421-80e0ec259618-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"87c34fde-d10d-4c46-a421-80e0ec259618\") " pod="openstack/ceilometer-0" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.385458 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87c34fde-d10d-4c46-a421-80e0ec259618-scripts\") pod \"ceilometer-0\" (UID: \"87c34fde-d10d-4c46-a421-80e0ec259618\") " pod="openstack/ceilometer-0" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.385477 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87c34fde-d10d-4c46-a421-80e0ec259618-config-data\") pod \"ceilometer-0\" (UID: \"87c34fde-d10d-4c46-a421-80e0ec259618\") " pod="openstack/ceilometer-0" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.385515 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q27rn\" (UniqueName: \"kubernetes.io/projected/87c34fde-d10d-4c46-a421-80e0ec259618-kube-api-access-q27rn\") pod \"ceilometer-0\" (UID: \"87c34fde-d10d-4c46-a421-80e0ec259618\") " pod="openstack/ceilometer-0" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.385539 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87c34fde-d10d-4c46-a421-80e0ec259618-run-httpd\") pod \"ceilometer-0\" (UID: \"87c34fde-d10d-4c46-a421-80e0ec259618\") " pod="openstack/ceilometer-0" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.386031 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87c34fde-d10d-4c46-a421-80e0ec259618-run-httpd\") pod \"ceilometer-0\" (UID: \"87c34fde-d10d-4c46-a421-80e0ec259618\") " pod="openstack/ceilometer-0" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.386195 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87c34fde-d10d-4c46-a421-80e0ec259618-log-httpd\") pod \"ceilometer-0\" (UID: \"87c34fde-d10d-4c46-a421-80e0ec259618\") " pod="openstack/ceilometer-0" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.391721 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/87c34fde-d10d-4c46-a421-80e0ec259618-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"87c34fde-d10d-4c46-a421-80e0ec259618\") " pod="openstack/ceilometer-0" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.391734 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c34fde-d10d-4c46-a421-80e0ec259618-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"87c34fde-d10d-4c46-a421-80e0ec259618\") " pod="openstack/ceilometer-0" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.391780 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87c34fde-d10d-4c46-a421-80e0ec259618-scripts\") pod \"ceilometer-0\" (UID: \"87c34fde-d10d-4c46-a421-80e0ec259618\") " pod="openstack/ceilometer-0" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.392525 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87c34fde-d10d-4c46-a421-80e0ec259618-config-data\") pod \"ceilometer-0\" (UID: \"87c34fde-d10d-4c46-a421-80e0ec259618\") " pod="openstack/ceilometer-0" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.392970 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87c34fde-d10d-4c46-a421-80e0ec259618-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"87c34fde-d10d-4c46-a421-80e0ec259618\") " pod="openstack/ceilometer-0" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.405663 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q27rn\" (UniqueName: \"kubernetes.io/projected/87c34fde-d10d-4c46-a421-80e0ec259618-kube-api-access-q27rn\") pod \"ceilometer-0\" (UID: \"87c34fde-d10d-4c46-a421-80e0ec259618\") " pod="openstack/ceilometer-0" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.426497 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:54:43 crc kubenswrapper[4787]: I1001 09:54:43.717519 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:54:43 crc kubenswrapper[4787]: W1001 09:54:43.719311 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87c34fde_d10d_4c46_a421_80e0ec259618.slice/crio-51e61818de9f4b31deb8939edd371f92f07d29cd0984d32557ca9b21406a98ad WatchSource:0}: Error finding container 51e61818de9f4b31deb8939edd371f92f07d29cd0984d32557ca9b21406a98ad: Status 404 returned error can't find the container with id 51e61818de9f4b31deb8939edd371f92f07d29cd0984d32557ca9b21406a98ad Oct 01 09:54:44 crc kubenswrapper[4787]: I1001 09:54:44.047327 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"87c34fde-d10d-4c46-a421-80e0ec259618","Type":"ContainerStarted","Data":"51e61818de9f4b31deb8939edd371f92f07d29cd0984d32557ca9b21406a98ad"} Oct 01 09:54:44 crc kubenswrapper[4787]: I1001 09:54:44.553060 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c67fad8-4bad-49ac-bd9f-218c6d0d98d5" path="/var/lib/kubelet/pods/5c67fad8-4bad-49ac-bd9f-218c6d0d98d5/volumes" Oct 01 09:54:45 crc kubenswrapper[4787]: I1001 09:54:45.059262 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"87c34fde-d10d-4c46-a421-80e0ec259618","Type":"ContainerStarted","Data":"8d6e18fe98ef9027ff8eb7eb0f6778f4bbc31504c85bea3b6ec35e8b1fda0c99"} Oct 01 09:54:45 crc kubenswrapper[4787]: I1001 09:54:45.380772 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 01 09:54:45 crc kubenswrapper[4787]: I1001 09:54:45.397966 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 01 09:54:45 crc kubenswrapper[4787]: I1001 09:54:45.398028 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 01 09:54:45 crc kubenswrapper[4787]: I1001 09:54:45.425634 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 01 09:54:46 crc kubenswrapper[4787]: I1001 09:54:46.071157 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"87c34fde-d10d-4c46-a421-80e0ec259618","Type":"ContainerStarted","Data":"9fa49d4036bc4bddff9f38920d3c3166c8c166f938fab2d7567edfcd765f035d"} Oct 01 09:54:46 crc kubenswrapper[4787]: I1001 09:54:46.071638 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"87c34fde-d10d-4c46-a421-80e0ec259618","Type":"ContainerStarted","Data":"2279e2f22be98966a586fefeac39a0f0d83c6396bcfce3feea7247405e00b973"} Oct 01 09:54:46 crc kubenswrapper[4787]: I1001 09:54:46.106315 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 01 09:54:46 crc kubenswrapper[4787]: I1001 09:54:46.480365 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ec425a80-032d-48b0-b46d-a38d98482149" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.195:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 01 09:54:46 crc kubenswrapper[4787]: I1001 09:54:46.480411 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ec425a80-032d-48b0-b46d-a38d98482149" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.195:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 01 09:54:48 crc kubenswrapper[4787]: I1001 09:54:48.107401 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"87c34fde-d10d-4c46-a421-80e0ec259618","Type":"ContainerStarted","Data":"9456b2948249c91faffcd699d3f7fec1eea21de1c78841b56f960beba8113270"} Oct 01 09:54:48 crc kubenswrapper[4787]: I1001 09:54:48.108235 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 01 09:54:48 crc kubenswrapper[4787]: I1001 09:54:48.145718 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.746080885 podStartE2EDuration="5.145678905s" podCreationTimestamp="2025-10-01 09:54:43 +0000 UTC" firstStartedPulling="2025-10-01 09:54:43.721362578 +0000 UTC m=+1115.836506735" lastFinishedPulling="2025-10-01 09:54:47.120960578 +0000 UTC m=+1119.236104755" observedRunningTime="2025-10-01 09:54:48.126661673 +0000 UTC m=+1120.241805840" watchObservedRunningTime="2025-10-01 09:54:48.145678905 +0000 UTC m=+1120.260823062" Oct 01 09:54:49 crc kubenswrapper[4787]: I1001 09:54:49.377047 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.159062 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.161910 4787 generic.go:334] "Generic (PLEG): container finished" podID="13182f8e-43d0-40f7-b5a4-f11f34998d3d" containerID="e38fa43fb47cc1a3b63c611be02241530a0bcf30baada4f0f1151c07e73f71eb" exitCode=137 Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.161978 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"13182f8e-43d0-40f7-b5a4-f11f34998d3d","Type":"ContainerDied","Data":"e38fa43fb47cc1a3b63c611be02241530a0bcf30baada4f0f1151c07e73f71eb"} Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.161996 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.162017 4787 scope.go:117] "RemoveContainer" containerID="e38fa43fb47cc1a3b63c611be02241530a0bcf30baada4f0f1151c07e73f71eb" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.162003 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"13182f8e-43d0-40f7-b5a4-f11f34998d3d","Type":"ContainerDied","Data":"2bd00980e99795af90e8253e592241aaeb05bf08b929df0b7c2bdf43b10afc37"} Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.235897 4787 generic.go:334] "Generic (PLEG): container finished" podID="2c7de75f-5388-4bda-97ef-d3a17234001b" containerID="50854a5e8554c38855622f13e3e07ceedacfd52a6adccf1ebbe564a1a31d2099" exitCode=137 Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.235942 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2c7de75f-5388-4bda-97ef-d3a17234001b","Type":"ContainerDied","Data":"50854a5e8554c38855622f13e3e07ceedacfd52a6adccf1ebbe564a1a31d2099"} Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.235971 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2c7de75f-5388-4bda-97ef-d3a17234001b","Type":"ContainerDied","Data":"2a93b0ceec1a8b3b37ff1f36dee2a606d62812014f2e640244b35aca3d3d07b9"} Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.235985 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a93b0ceec1a8b3b37ff1f36dee2a606d62812014f2e640244b35aca3d3d07b9" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.241408 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.266041 4787 scope.go:117] "RemoveContainer" containerID="e38fa43fb47cc1a3b63c611be02241530a0bcf30baada4f0f1151c07e73f71eb" Oct 01 09:54:53 crc kubenswrapper[4787]: E1001 09:54:53.266546 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e38fa43fb47cc1a3b63c611be02241530a0bcf30baada4f0f1151c07e73f71eb\": container with ID starting with e38fa43fb47cc1a3b63c611be02241530a0bcf30baada4f0f1151c07e73f71eb not found: ID does not exist" containerID="e38fa43fb47cc1a3b63c611be02241530a0bcf30baada4f0f1151c07e73f71eb" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.266580 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e38fa43fb47cc1a3b63c611be02241530a0bcf30baada4f0f1151c07e73f71eb"} err="failed to get container status \"e38fa43fb47cc1a3b63c611be02241530a0bcf30baada4f0f1151c07e73f71eb\": rpc error: code = NotFound desc = could not find container \"e38fa43fb47cc1a3b63c611be02241530a0bcf30baada4f0f1151c07e73f71eb\": container with ID starting with e38fa43fb47cc1a3b63c611be02241530a0bcf30baada4f0f1151c07e73f71eb not found: ID does not exist" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.327361 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13182f8e-43d0-40f7-b5a4-f11f34998d3d-combined-ca-bundle\") pod \"13182f8e-43d0-40f7-b5a4-f11f34998d3d\" (UID: \"13182f8e-43d0-40f7-b5a4-f11f34998d3d\") " Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.327416 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54fpx\" (UniqueName: \"kubernetes.io/projected/13182f8e-43d0-40f7-b5a4-f11f34998d3d-kube-api-access-54fpx\") pod \"13182f8e-43d0-40f7-b5a4-f11f34998d3d\" (UID: \"13182f8e-43d0-40f7-b5a4-f11f34998d3d\") " Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.327476 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13182f8e-43d0-40f7-b5a4-f11f34998d3d-config-data\") pod \"13182f8e-43d0-40f7-b5a4-f11f34998d3d\" (UID: \"13182f8e-43d0-40f7-b5a4-f11f34998d3d\") " Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.334263 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13182f8e-43d0-40f7-b5a4-f11f34998d3d-kube-api-access-54fpx" (OuterVolumeSpecName: "kube-api-access-54fpx") pod "13182f8e-43d0-40f7-b5a4-f11f34998d3d" (UID: "13182f8e-43d0-40f7-b5a4-f11f34998d3d"). InnerVolumeSpecName "kube-api-access-54fpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.355091 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13182f8e-43d0-40f7-b5a4-f11f34998d3d-config-data" (OuterVolumeSpecName: "config-data") pod "13182f8e-43d0-40f7-b5a4-f11f34998d3d" (UID: "13182f8e-43d0-40f7-b5a4-f11f34998d3d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.355489 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13182f8e-43d0-40f7-b5a4-f11f34998d3d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "13182f8e-43d0-40f7-b5a4-f11f34998d3d" (UID: "13182f8e-43d0-40f7-b5a4-f11f34998d3d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.428528 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c7de75f-5388-4bda-97ef-d3a17234001b-logs\") pod \"2c7de75f-5388-4bda-97ef-d3a17234001b\" (UID: \"2c7de75f-5388-4bda-97ef-d3a17234001b\") " Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.428638 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ccmqn\" (UniqueName: \"kubernetes.io/projected/2c7de75f-5388-4bda-97ef-d3a17234001b-kube-api-access-ccmqn\") pod \"2c7de75f-5388-4bda-97ef-d3a17234001b\" (UID: \"2c7de75f-5388-4bda-97ef-d3a17234001b\") " Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.428694 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c7de75f-5388-4bda-97ef-d3a17234001b-config-data\") pod \"2c7de75f-5388-4bda-97ef-d3a17234001b\" (UID: \"2c7de75f-5388-4bda-97ef-d3a17234001b\") " Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.428841 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c7de75f-5388-4bda-97ef-d3a17234001b-combined-ca-bundle\") pod \"2c7de75f-5388-4bda-97ef-d3a17234001b\" (UID: \"2c7de75f-5388-4bda-97ef-d3a17234001b\") " Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.429393 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13182f8e-43d0-40f7-b5a4-f11f34998d3d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.429412 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54fpx\" (UniqueName: \"kubernetes.io/projected/13182f8e-43d0-40f7-b5a4-f11f34998d3d-kube-api-access-54fpx\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.429424 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13182f8e-43d0-40f7-b5a4-f11f34998d3d-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.430064 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c7de75f-5388-4bda-97ef-d3a17234001b-logs" (OuterVolumeSpecName: "logs") pod "2c7de75f-5388-4bda-97ef-d3a17234001b" (UID: "2c7de75f-5388-4bda-97ef-d3a17234001b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.436678 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c7de75f-5388-4bda-97ef-d3a17234001b-kube-api-access-ccmqn" (OuterVolumeSpecName: "kube-api-access-ccmqn") pod "2c7de75f-5388-4bda-97ef-d3a17234001b" (UID: "2c7de75f-5388-4bda-97ef-d3a17234001b"). InnerVolumeSpecName "kube-api-access-ccmqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.454427 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c7de75f-5388-4bda-97ef-d3a17234001b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2c7de75f-5388-4bda-97ef-d3a17234001b" (UID: "2c7de75f-5388-4bda-97ef-d3a17234001b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.459010 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c7de75f-5388-4bda-97ef-d3a17234001b-config-data" (OuterVolumeSpecName: "config-data") pod "2c7de75f-5388-4bda-97ef-d3a17234001b" (UID: "2c7de75f-5388-4bda-97ef-d3a17234001b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.494788 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.506357 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.515735 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 09:54:53 crc kubenswrapper[4787]: E1001 09:54:53.516391 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c7de75f-5388-4bda-97ef-d3a17234001b" containerName="nova-metadata-metadata" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.516463 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c7de75f-5388-4bda-97ef-d3a17234001b" containerName="nova-metadata-metadata" Oct 01 09:54:53 crc kubenswrapper[4787]: E1001 09:54:53.516540 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c7de75f-5388-4bda-97ef-d3a17234001b" containerName="nova-metadata-log" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.516600 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c7de75f-5388-4bda-97ef-d3a17234001b" containerName="nova-metadata-log" Oct 01 09:54:53 crc kubenswrapper[4787]: E1001 09:54:53.516674 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13182f8e-43d0-40f7-b5a4-f11f34998d3d" containerName="nova-cell1-novncproxy-novncproxy" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.516735 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="13182f8e-43d0-40f7-b5a4-f11f34998d3d" containerName="nova-cell1-novncproxy-novncproxy" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.516962 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c7de75f-5388-4bda-97ef-d3a17234001b" containerName="nova-metadata-log" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.517101 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="13182f8e-43d0-40f7-b5a4-f11f34998d3d" containerName="nova-cell1-novncproxy-novncproxy" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.517171 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c7de75f-5388-4bda-97ef-d3a17234001b" containerName="nova-metadata-metadata" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.517819 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.521311 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.521720 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.522367 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.531127 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c7de75f-5388-4bda-97ef-d3a17234001b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.531159 4787 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c7de75f-5388-4bda-97ef-d3a17234001b-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.531169 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ccmqn\" (UniqueName: \"kubernetes.io/projected/2c7de75f-5388-4bda-97ef-d3a17234001b-kube-api-access-ccmqn\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.531179 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c7de75f-5388-4bda-97ef-d3a17234001b-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.536372 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.633194 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b58hw\" (UniqueName: \"kubernetes.io/projected/f8b719ce-ceea-4c80-8c8d-846f2bf7a402-kube-api-access-b58hw\") pod \"nova-cell1-novncproxy-0\" (UID: \"f8b719ce-ceea-4c80-8c8d-846f2bf7a402\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.633881 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8b719ce-ceea-4c80-8c8d-846f2bf7a402-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"f8b719ce-ceea-4c80-8c8d-846f2bf7a402\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.633999 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8b719ce-ceea-4c80-8c8d-846f2bf7a402-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"f8b719ce-ceea-4c80-8c8d-846f2bf7a402\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.634101 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8b719ce-ceea-4c80-8c8d-846f2bf7a402-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f8b719ce-ceea-4c80-8c8d-846f2bf7a402\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.634225 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8b719ce-ceea-4c80-8c8d-846f2bf7a402-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f8b719ce-ceea-4c80-8c8d-846f2bf7a402\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.737499 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8b719ce-ceea-4c80-8c8d-846f2bf7a402-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"f8b719ce-ceea-4c80-8c8d-846f2bf7a402\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.737557 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8b719ce-ceea-4c80-8c8d-846f2bf7a402-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f8b719ce-ceea-4c80-8c8d-846f2bf7a402\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.737630 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8b719ce-ceea-4c80-8c8d-846f2bf7a402-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f8b719ce-ceea-4c80-8c8d-846f2bf7a402\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.737686 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b58hw\" (UniqueName: \"kubernetes.io/projected/f8b719ce-ceea-4c80-8c8d-846f2bf7a402-kube-api-access-b58hw\") pod \"nova-cell1-novncproxy-0\" (UID: \"f8b719ce-ceea-4c80-8c8d-846f2bf7a402\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.737730 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8b719ce-ceea-4c80-8c8d-846f2bf7a402-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"f8b719ce-ceea-4c80-8c8d-846f2bf7a402\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.743013 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8b719ce-ceea-4c80-8c8d-846f2bf7a402-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"f8b719ce-ceea-4c80-8c8d-846f2bf7a402\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.743159 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8b719ce-ceea-4c80-8c8d-846f2bf7a402-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f8b719ce-ceea-4c80-8c8d-846f2bf7a402\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.744948 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8b719ce-ceea-4c80-8c8d-846f2bf7a402-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"f8b719ce-ceea-4c80-8c8d-846f2bf7a402\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.750320 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8b719ce-ceea-4c80-8c8d-846f2bf7a402-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f8b719ce-ceea-4c80-8c8d-846f2bf7a402\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.758224 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b58hw\" (UniqueName: \"kubernetes.io/projected/f8b719ce-ceea-4c80-8c8d-846f2bf7a402-kube-api-access-b58hw\") pod \"nova-cell1-novncproxy-0\" (UID: \"f8b719ce-ceea-4c80-8c8d-846f2bf7a402\") " pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:54:53 crc kubenswrapper[4787]: I1001 09:54:53.850683 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:54:54 crc kubenswrapper[4787]: I1001 09:54:54.253202 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 09:54:54 crc kubenswrapper[4787]: I1001 09:54:54.299806 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:54:54 crc kubenswrapper[4787]: I1001 09:54:54.309899 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:54:54 crc kubenswrapper[4787]: I1001 09:54:54.329825 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:54:54 crc kubenswrapper[4787]: I1001 09:54:54.332652 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 09:54:54 crc kubenswrapper[4787]: I1001 09:54:54.337298 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 01 09:54:54 crc kubenswrapper[4787]: I1001 09:54:54.337625 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 01 09:54:54 crc kubenswrapper[4787]: I1001 09:54:54.347278 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:54:54 crc kubenswrapper[4787]: I1001 09:54:54.354707 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/700e3732-7195-48fb-a5f6-4fb80b9f13e8-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"700e3732-7195-48fb-a5f6-4fb80b9f13e8\") " pod="openstack/nova-metadata-0" Oct 01 09:54:54 crc kubenswrapper[4787]: I1001 09:54:54.354824 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzd4v\" (UniqueName: \"kubernetes.io/projected/700e3732-7195-48fb-a5f6-4fb80b9f13e8-kube-api-access-hzd4v\") pod \"nova-metadata-0\" (UID: \"700e3732-7195-48fb-a5f6-4fb80b9f13e8\") " pod="openstack/nova-metadata-0" Oct 01 09:54:54 crc kubenswrapper[4787]: I1001 09:54:54.354936 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/700e3732-7195-48fb-a5f6-4fb80b9f13e8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"700e3732-7195-48fb-a5f6-4fb80b9f13e8\") " pod="openstack/nova-metadata-0" Oct 01 09:54:54 crc kubenswrapper[4787]: I1001 09:54:54.355052 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/700e3732-7195-48fb-a5f6-4fb80b9f13e8-logs\") pod \"nova-metadata-0\" (UID: \"700e3732-7195-48fb-a5f6-4fb80b9f13e8\") " pod="openstack/nova-metadata-0" Oct 01 09:54:54 crc kubenswrapper[4787]: I1001 09:54:54.355118 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/700e3732-7195-48fb-a5f6-4fb80b9f13e8-config-data\") pod \"nova-metadata-0\" (UID: \"700e3732-7195-48fb-a5f6-4fb80b9f13e8\") " pod="openstack/nova-metadata-0" Oct 01 09:54:54 crc kubenswrapper[4787]: I1001 09:54:54.385950 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 01 09:54:54 crc kubenswrapper[4787]: W1001 09:54:54.386443 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf8b719ce_ceea_4c80_8c8d_846f2bf7a402.slice/crio-58625957b28e881db8e411a49673436a6e70b871482607c49337b704484d6faa WatchSource:0}: Error finding container 58625957b28e881db8e411a49673436a6e70b871482607c49337b704484d6faa: Status 404 returned error can't find the container with id 58625957b28e881db8e411a49673436a6e70b871482607c49337b704484d6faa Oct 01 09:54:54 crc kubenswrapper[4787]: I1001 09:54:54.456385 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzd4v\" (UniqueName: \"kubernetes.io/projected/700e3732-7195-48fb-a5f6-4fb80b9f13e8-kube-api-access-hzd4v\") pod \"nova-metadata-0\" (UID: \"700e3732-7195-48fb-a5f6-4fb80b9f13e8\") " pod="openstack/nova-metadata-0" Oct 01 09:54:54 crc kubenswrapper[4787]: I1001 09:54:54.456461 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/700e3732-7195-48fb-a5f6-4fb80b9f13e8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"700e3732-7195-48fb-a5f6-4fb80b9f13e8\") " pod="openstack/nova-metadata-0" Oct 01 09:54:54 crc kubenswrapper[4787]: I1001 09:54:54.456513 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/700e3732-7195-48fb-a5f6-4fb80b9f13e8-logs\") pod \"nova-metadata-0\" (UID: \"700e3732-7195-48fb-a5f6-4fb80b9f13e8\") " pod="openstack/nova-metadata-0" Oct 01 09:54:54 crc kubenswrapper[4787]: I1001 09:54:54.456541 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/700e3732-7195-48fb-a5f6-4fb80b9f13e8-config-data\") pod \"nova-metadata-0\" (UID: \"700e3732-7195-48fb-a5f6-4fb80b9f13e8\") " pod="openstack/nova-metadata-0" Oct 01 09:54:54 crc kubenswrapper[4787]: I1001 09:54:54.456606 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/700e3732-7195-48fb-a5f6-4fb80b9f13e8-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"700e3732-7195-48fb-a5f6-4fb80b9f13e8\") " pod="openstack/nova-metadata-0" Oct 01 09:54:54 crc kubenswrapper[4787]: I1001 09:54:54.457666 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/700e3732-7195-48fb-a5f6-4fb80b9f13e8-logs\") pod \"nova-metadata-0\" (UID: \"700e3732-7195-48fb-a5f6-4fb80b9f13e8\") " pod="openstack/nova-metadata-0" Oct 01 09:54:54 crc kubenswrapper[4787]: I1001 09:54:54.462666 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/700e3732-7195-48fb-a5f6-4fb80b9f13e8-config-data\") pod \"nova-metadata-0\" (UID: \"700e3732-7195-48fb-a5f6-4fb80b9f13e8\") " pod="openstack/nova-metadata-0" Oct 01 09:54:54 crc kubenswrapper[4787]: I1001 09:54:54.462701 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/700e3732-7195-48fb-a5f6-4fb80b9f13e8-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"700e3732-7195-48fb-a5f6-4fb80b9f13e8\") " pod="openstack/nova-metadata-0" Oct 01 09:54:54 crc kubenswrapper[4787]: I1001 09:54:54.463925 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/700e3732-7195-48fb-a5f6-4fb80b9f13e8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"700e3732-7195-48fb-a5f6-4fb80b9f13e8\") " pod="openstack/nova-metadata-0" Oct 01 09:54:54 crc kubenswrapper[4787]: I1001 09:54:54.477036 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzd4v\" (UniqueName: \"kubernetes.io/projected/700e3732-7195-48fb-a5f6-4fb80b9f13e8-kube-api-access-hzd4v\") pod \"nova-metadata-0\" (UID: \"700e3732-7195-48fb-a5f6-4fb80b9f13e8\") " pod="openstack/nova-metadata-0" Oct 01 09:54:54 crc kubenswrapper[4787]: I1001 09:54:54.538092 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13182f8e-43d0-40f7-b5a4-f11f34998d3d" path="/var/lib/kubelet/pods/13182f8e-43d0-40f7-b5a4-f11f34998d3d/volumes" Oct 01 09:54:54 crc kubenswrapper[4787]: I1001 09:54:54.539159 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c7de75f-5388-4bda-97ef-d3a17234001b" path="/var/lib/kubelet/pods/2c7de75f-5388-4bda-97ef-d3a17234001b/volumes" Oct 01 09:54:54 crc kubenswrapper[4787]: I1001 09:54:54.670107 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 09:54:54 crc kubenswrapper[4787]: I1001 09:54:54.934443 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:54:54 crc kubenswrapper[4787]: W1001 09:54:54.941326 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod700e3732_7195_48fb_a5f6_4fb80b9f13e8.slice/crio-08d31b9e424648c555b249db62424526da244f58e3101e65ff6d91d738ad8729 WatchSource:0}: Error finding container 08d31b9e424648c555b249db62424526da244f58e3101e65ff6d91d738ad8729: Status 404 returned error can't find the container with id 08d31b9e424648c555b249db62424526da244f58e3101e65ff6d91d738ad8729 Oct 01 09:54:55 crc kubenswrapper[4787]: I1001 09:54:55.275504 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"700e3732-7195-48fb-a5f6-4fb80b9f13e8","Type":"ContainerStarted","Data":"5a17511d23a3c2c840e42020b72de1cd5076c84cd25e857681d078bc8332adca"} Oct 01 09:54:55 crc kubenswrapper[4787]: I1001 09:54:55.275594 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"700e3732-7195-48fb-a5f6-4fb80b9f13e8","Type":"ContainerStarted","Data":"08d31b9e424648c555b249db62424526da244f58e3101e65ff6d91d738ad8729"} Oct 01 09:54:55 crc kubenswrapper[4787]: I1001 09:54:55.278431 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"f8b719ce-ceea-4c80-8c8d-846f2bf7a402","Type":"ContainerStarted","Data":"365b6052addc6af7164c03abf7c8267c7b9b44a79d91c3dc9729f7a1f58b8175"} Oct 01 09:54:55 crc kubenswrapper[4787]: I1001 09:54:55.278483 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"f8b719ce-ceea-4c80-8c8d-846f2bf7a402","Type":"ContainerStarted","Data":"58625957b28e881db8e411a49673436a6e70b871482607c49337b704484d6faa"} Oct 01 09:54:55 crc kubenswrapper[4787]: I1001 09:54:55.309959 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.309911491 podStartE2EDuration="2.309911491s" podCreationTimestamp="2025-10-01 09:54:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:54:55.306912505 +0000 UTC m=+1127.422056732" watchObservedRunningTime="2025-10-01 09:54:55.309911491 +0000 UTC m=+1127.425055668" Oct 01 09:54:55 crc kubenswrapper[4787]: I1001 09:54:55.402584 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 01 09:54:55 crc kubenswrapper[4787]: I1001 09:54:55.403003 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 01 09:54:55 crc kubenswrapper[4787]: I1001 09:54:55.406033 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 01 09:54:55 crc kubenswrapper[4787]: I1001 09:54:55.415724 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 01 09:54:56 crc kubenswrapper[4787]: I1001 09:54:56.292221 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"700e3732-7195-48fb-a5f6-4fb80b9f13e8","Type":"ContainerStarted","Data":"9a4e6b2ebf2a9aee84a3e28a27294b0ceba00c22a18adce42fe141eb75395696"} Oct 01 09:54:56 crc kubenswrapper[4787]: I1001 09:54:56.293773 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 01 09:54:56 crc kubenswrapper[4787]: I1001 09:54:56.297153 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 01 09:54:56 crc kubenswrapper[4787]: I1001 09:54:56.325706 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.325682481 podStartE2EDuration="2.325682481s" podCreationTimestamp="2025-10-01 09:54:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:54:56.320150571 +0000 UTC m=+1128.435294748" watchObservedRunningTime="2025-10-01 09:54:56.325682481 +0000 UTC m=+1128.440826638" Oct 01 09:54:56 crc kubenswrapper[4787]: I1001 09:54:56.540602 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-68b59d98cf-xk49b"] Oct 01 09:54:56 crc kubenswrapper[4787]: I1001 09:54:56.542258 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68b59d98cf-xk49b" Oct 01 09:54:56 crc kubenswrapper[4787]: I1001 09:54:56.554301 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68b59d98cf-xk49b"] Oct 01 09:54:56 crc kubenswrapper[4787]: I1001 09:54:56.715767 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ed8acab0-e099-4b65-9883-5ea45cc4599e-dns-swift-storage-0\") pod \"dnsmasq-dns-68b59d98cf-xk49b\" (UID: \"ed8acab0-e099-4b65-9883-5ea45cc4599e\") " pod="openstack/dnsmasq-dns-68b59d98cf-xk49b" Oct 01 09:54:56 crc kubenswrapper[4787]: I1001 09:54:56.717557 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed8acab0-e099-4b65-9883-5ea45cc4599e-config\") pod \"dnsmasq-dns-68b59d98cf-xk49b\" (UID: \"ed8acab0-e099-4b65-9883-5ea45cc4599e\") " pod="openstack/dnsmasq-dns-68b59d98cf-xk49b" Oct 01 09:54:56 crc kubenswrapper[4787]: I1001 09:54:56.717654 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvk7q\" (UniqueName: \"kubernetes.io/projected/ed8acab0-e099-4b65-9883-5ea45cc4599e-kube-api-access-dvk7q\") pod \"dnsmasq-dns-68b59d98cf-xk49b\" (UID: \"ed8acab0-e099-4b65-9883-5ea45cc4599e\") " pod="openstack/dnsmasq-dns-68b59d98cf-xk49b" Oct 01 09:54:56 crc kubenswrapper[4787]: I1001 09:54:56.717759 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed8acab0-e099-4b65-9883-5ea45cc4599e-dns-svc\") pod \"dnsmasq-dns-68b59d98cf-xk49b\" (UID: \"ed8acab0-e099-4b65-9883-5ea45cc4599e\") " pod="openstack/dnsmasq-dns-68b59d98cf-xk49b" Oct 01 09:54:56 crc kubenswrapper[4787]: I1001 09:54:56.717832 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed8acab0-e099-4b65-9883-5ea45cc4599e-ovsdbserver-sb\") pod \"dnsmasq-dns-68b59d98cf-xk49b\" (UID: \"ed8acab0-e099-4b65-9883-5ea45cc4599e\") " pod="openstack/dnsmasq-dns-68b59d98cf-xk49b" Oct 01 09:54:56 crc kubenswrapper[4787]: I1001 09:54:56.717863 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed8acab0-e099-4b65-9883-5ea45cc4599e-ovsdbserver-nb\") pod \"dnsmasq-dns-68b59d98cf-xk49b\" (UID: \"ed8acab0-e099-4b65-9883-5ea45cc4599e\") " pod="openstack/dnsmasq-dns-68b59d98cf-xk49b" Oct 01 09:54:56 crc kubenswrapper[4787]: I1001 09:54:56.820387 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ed8acab0-e099-4b65-9883-5ea45cc4599e-dns-swift-storage-0\") pod \"dnsmasq-dns-68b59d98cf-xk49b\" (UID: \"ed8acab0-e099-4b65-9883-5ea45cc4599e\") " pod="openstack/dnsmasq-dns-68b59d98cf-xk49b" Oct 01 09:54:56 crc kubenswrapper[4787]: I1001 09:54:56.820612 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed8acab0-e099-4b65-9883-5ea45cc4599e-config\") pod \"dnsmasq-dns-68b59d98cf-xk49b\" (UID: \"ed8acab0-e099-4b65-9883-5ea45cc4599e\") " pod="openstack/dnsmasq-dns-68b59d98cf-xk49b" Oct 01 09:54:56 crc kubenswrapper[4787]: I1001 09:54:56.820703 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvk7q\" (UniqueName: \"kubernetes.io/projected/ed8acab0-e099-4b65-9883-5ea45cc4599e-kube-api-access-dvk7q\") pod \"dnsmasq-dns-68b59d98cf-xk49b\" (UID: \"ed8acab0-e099-4b65-9883-5ea45cc4599e\") " pod="openstack/dnsmasq-dns-68b59d98cf-xk49b" Oct 01 09:54:56 crc kubenswrapper[4787]: I1001 09:54:56.820801 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed8acab0-e099-4b65-9883-5ea45cc4599e-dns-svc\") pod \"dnsmasq-dns-68b59d98cf-xk49b\" (UID: \"ed8acab0-e099-4b65-9883-5ea45cc4599e\") " pod="openstack/dnsmasq-dns-68b59d98cf-xk49b" Oct 01 09:54:56 crc kubenswrapper[4787]: I1001 09:54:56.820866 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed8acab0-e099-4b65-9883-5ea45cc4599e-ovsdbserver-sb\") pod \"dnsmasq-dns-68b59d98cf-xk49b\" (UID: \"ed8acab0-e099-4b65-9883-5ea45cc4599e\") " pod="openstack/dnsmasq-dns-68b59d98cf-xk49b" Oct 01 09:54:56 crc kubenswrapper[4787]: I1001 09:54:56.820910 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed8acab0-e099-4b65-9883-5ea45cc4599e-ovsdbserver-nb\") pod \"dnsmasq-dns-68b59d98cf-xk49b\" (UID: \"ed8acab0-e099-4b65-9883-5ea45cc4599e\") " pod="openstack/dnsmasq-dns-68b59d98cf-xk49b" Oct 01 09:54:56 crc kubenswrapper[4787]: I1001 09:54:56.822718 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed8acab0-e099-4b65-9883-5ea45cc4599e-ovsdbserver-nb\") pod \"dnsmasq-dns-68b59d98cf-xk49b\" (UID: \"ed8acab0-e099-4b65-9883-5ea45cc4599e\") " pod="openstack/dnsmasq-dns-68b59d98cf-xk49b" Oct 01 09:54:56 crc kubenswrapper[4787]: I1001 09:54:56.824702 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed8acab0-e099-4b65-9883-5ea45cc4599e-config\") pod \"dnsmasq-dns-68b59d98cf-xk49b\" (UID: \"ed8acab0-e099-4b65-9883-5ea45cc4599e\") " pod="openstack/dnsmasq-dns-68b59d98cf-xk49b" Oct 01 09:54:56 crc kubenswrapper[4787]: I1001 09:54:56.825209 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ed8acab0-e099-4b65-9883-5ea45cc4599e-dns-swift-storage-0\") pod \"dnsmasq-dns-68b59d98cf-xk49b\" (UID: \"ed8acab0-e099-4b65-9883-5ea45cc4599e\") " pod="openstack/dnsmasq-dns-68b59d98cf-xk49b" Oct 01 09:54:56 crc kubenswrapper[4787]: I1001 09:54:56.825882 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed8acab0-e099-4b65-9883-5ea45cc4599e-dns-svc\") pod \"dnsmasq-dns-68b59d98cf-xk49b\" (UID: \"ed8acab0-e099-4b65-9883-5ea45cc4599e\") " pod="openstack/dnsmasq-dns-68b59d98cf-xk49b" Oct 01 09:54:56 crc kubenswrapper[4787]: I1001 09:54:56.826696 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed8acab0-e099-4b65-9883-5ea45cc4599e-ovsdbserver-sb\") pod \"dnsmasq-dns-68b59d98cf-xk49b\" (UID: \"ed8acab0-e099-4b65-9883-5ea45cc4599e\") " pod="openstack/dnsmasq-dns-68b59d98cf-xk49b" Oct 01 09:54:56 crc kubenswrapper[4787]: I1001 09:54:56.853785 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvk7q\" (UniqueName: \"kubernetes.io/projected/ed8acab0-e099-4b65-9883-5ea45cc4599e-kube-api-access-dvk7q\") pod \"dnsmasq-dns-68b59d98cf-xk49b\" (UID: \"ed8acab0-e099-4b65-9883-5ea45cc4599e\") " pod="openstack/dnsmasq-dns-68b59d98cf-xk49b" Oct 01 09:54:56 crc kubenswrapper[4787]: I1001 09:54:56.866142 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68b59d98cf-xk49b" Oct 01 09:54:57 crc kubenswrapper[4787]: I1001 09:54:57.382132 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68b59d98cf-xk49b"] Oct 01 09:54:58 crc kubenswrapper[4787]: I1001 09:54:58.325307 4787 generic.go:334] "Generic (PLEG): container finished" podID="ed8acab0-e099-4b65-9883-5ea45cc4599e" containerID="f1bc5cf1281c1f641d7c326c41bd88fcd42be40d931224664f09b2e9338df0cb" exitCode=0 Oct 01 09:54:58 crc kubenswrapper[4787]: I1001 09:54:58.325423 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68b59d98cf-xk49b" event={"ID":"ed8acab0-e099-4b65-9883-5ea45cc4599e","Type":"ContainerDied","Data":"f1bc5cf1281c1f641d7c326c41bd88fcd42be40d931224664f09b2e9338df0cb"} Oct 01 09:54:58 crc kubenswrapper[4787]: I1001 09:54:58.325493 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68b59d98cf-xk49b" event={"ID":"ed8acab0-e099-4b65-9883-5ea45cc4599e","Type":"ContainerStarted","Data":"6826e6bae9ecd4cbe77a334450c216d02fcc72a1432592f33335bd0a990ce6ec"} Oct 01 09:54:58 crc kubenswrapper[4787]: I1001 09:54:58.555528 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:54:58 crc kubenswrapper[4787]: I1001 09:54:58.558005 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="87c34fde-d10d-4c46-a421-80e0ec259618" containerName="ceilometer-central-agent" containerID="cri-o://8d6e18fe98ef9027ff8eb7eb0f6778f4bbc31504c85bea3b6ec35e8b1fda0c99" gracePeriod=30 Oct 01 09:54:58 crc kubenswrapper[4787]: I1001 09:54:58.558198 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="87c34fde-d10d-4c46-a421-80e0ec259618" containerName="sg-core" containerID="cri-o://9fa49d4036bc4bddff9f38920d3c3166c8c166f938fab2d7567edfcd765f035d" gracePeriod=30 Oct 01 09:54:58 crc kubenswrapper[4787]: I1001 09:54:58.558192 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="87c34fde-d10d-4c46-a421-80e0ec259618" containerName="proxy-httpd" containerID="cri-o://9456b2948249c91faffcd699d3f7fec1eea21de1c78841b56f960beba8113270" gracePeriod=30 Oct 01 09:54:58 crc kubenswrapper[4787]: I1001 09:54:58.558293 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="87c34fde-d10d-4c46-a421-80e0ec259618" containerName="ceilometer-notification-agent" containerID="cri-o://2279e2f22be98966a586fefeac39a0f0d83c6396bcfce3feea7247405e00b973" gracePeriod=30 Oct 01 09:54:58 crc kubenswrapper[4787]: I1001 09:54:58.584284 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="87c34fde-d10d-4c46-a421-80e0ec259618" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.198:3000/\": read tcp 10.217.0.2:33862->10.217.0.198:3000: read: connection reset by peer" Oct 01 09:54:58 crc kubenswrapper[4787]: I1001 09:54:58.850960 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:54:59 crc kubenswrapper[4787]: I1001 09:54:59.233855 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 01 09:54:59 crc kubenswrapper[4787]: I1001 09:54:59.339245 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68b59d98cf-xk49b" event={"ID":"ed8acab0-e099-4b65-9883-5ea45cc4599e","Type":"ContainerStarted","Data":"3bf412fd2bd6296d4602149f7c73e11ccd49430d8c5803653fcadeab6637bfa5"} Oct 01 09:54:59 crc kubenswrapper[4787]: I1001 09:54:59.339487 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-68b59d98cf-xk49b" Oct 01 09:54:59 crc kubenswrapper[4787]: I1001 09:54:59.342057 4787 generic.go:334] "Generic (PLEG): container finished" podID="87c34fde-d10d-4c46-a421-80e0ec259618" containerID="9456b2948249c91faffcd699d3f7fec1eea21de1c78841b56f960beba8113270" exitCode=0 Oct 01 09:54:59 crc kubenswrapper[4787]: I1001 09:54:59.342111 4787 generic.go:334] "Generic (PLEG): container finished" podID="87c34fde-d10d-4c46-a421-80e0ec259618" containerID="9fa49d4036bc4bddff9f38920d3c3166c8c166f938fab2d7567edfcd765f035d" exitCode=2 Oct 01 09:54:59 crc kubenswrapper[4787]: I1001 09:54:59.342126 4787 generic.go:334] "Generic (PLEG): container finished" podID="87c34fde-d10d-4c46-a421-80e0ec259618" containerID="8d6e18fe98ef9027ff8eb7eb0f6778f4bbc31504c85bea3b6ec35e8b1fda0c99" exitCode=0 Oct 01 09:54:59 crc kubenswrapper[4787]: I1001 09:54:59.342100 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"87c34fde-d10d-4c46-a421-80e0ec259618","Type":"ContainerDied","Data":"9456b2948249c91faffcd699d3f7fec1eea21de1c78841b56f960beba8113270"} Oct 01 09:54:59 crc kubenswrapper[4787]: I1001 09:54:59.342217 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"87c34fde-d10d-4c46-a421-80e0ec259618","Type":"ContainerDied","Data":"9fa49d4036bc4bddff9f38920d3c3166c8c166f938fab2d7567edfcd765f035d"} Oct 01 09:54:59 crc kubenswrapper[4787]: I1001 09:54:59.342234 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"87c34fde-d10d-4c46-a421-80e0ec259618","Type":"ContainerDied","Data":"8d6e18fe98ef9027ff8eb7eb0f6778f4bbc31504c85bea3b6ec35e8b1fda0c99"} Oct 01 09:54:59 crc kubenswrapper[4787]: I1001 09:54:59.342305 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ec425a80-032d-48b0-b46d-a38d98482149" containerName="nova-api-log" containerID="cri-o://aac3eccdc8574c073088fc9f79a2eb36d2a4b971b79fead4de1bf5723a6378a6" gracePeriod=30 Oct 01 09:54:59 crc kubenswrapper[4787]: I1001 09:54:59.342370 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ec425a80-032d-48b0-b46d-a38d98482149" containerName="nova-api-api" containerID="cri-o://47877f416671132ec65ad4daf6c9681db228d733d74960a9216ae06960eed683" gracePeriod=30 Oct 01 09:54:59 crc kubenswrapper[4787]: I1001 09:54:59.371066 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-68b59d98cf-xk49b" podStartSLOduration=3.371043002 podStartE2EDuration="3.371043002s" podCreationTimestamp="2025-10-01 09:54:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:54:59.361065028 +0000 UTC m=+1131.476209185" watchObservedRunningTime="2025-10-01 09:54:59.371043002 +0000 UTC m=+1131.486187159" Oct 01 09:54:59 crc kubenswrapper[4787]: I1001 09:54:59.670639 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 01 09:54:59 crc kubenswrapper[4787]: I1001 09:54:59.670971 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 01 09:55:00 crc kubenswrapper[4787]: I1001 09:55:00.353581 4787 generic.go:334] "Generic (PLEG): container finished" podID="ec425a80-032d-48b0-b46d-a38d98482149" containerID="aac3eccdc8574c073088fc9f79a2eb36d2a4b971b79fead4de1bf5723a6378a6" exitCode=143 Oct 01 09:55:00 crc kubenswrapper[4787]: I1001 09:55:00.354599 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ec425a80-032d-48b0-b46d-a38d98482149","Type":"ContainerDied","Data":"aac3eccdc8574c073088fc9f79a2eb36d2a4b971b79fead4de1bf5723a6378a6"} Oct 01 09:55:01 crc kubenswrapper[4787]: I1001 09:55:01.366503 4787 generic.go:334] "Generic (PLEG): container finished" podID="87c34fde-d10d-4c46-a421-80e0ec259618" containerID="2279e2f22be98966a586fefeac39a0f0d83c6396bcfce3feea7247405e00b973" exitCode=0 Oct 01 09:55:01 crc kubenswrapper[4787]: I1001 09:55:01.366547 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"87c34fde-d10d-4c46-a421-80e0ec259618","Type":"ContainerDied","Data":"2279e2f22be98966a586fefeac39a0f0d83c6396bcfce3feea7247405e00b973"} Oct 01 09:55:01 crc kubenswrapper[4787]: I1001 09:55:01.632868 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:55:01 crc kubenswrapper[4787]: I1001 09:55:01.646422 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q27rn\" (UniqueName: \"kubernetes.io/projected/87c34fde-d10d-4c46-a421-80e0ec259618-kube-api-access-q27rn\") pod \"87c34fde-d10d-4c46-a421-80e0ec259618\" (UID: \"87c34fde-d10d-4c46-a421-80e0ec259618\") " Oct 01 09:55:01 crc kubenswrapper[4787]: I1001 09:55:01.646695 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/87c34fde-d10d-4c46-a421-80e0ec259618-sg-core-conf-yaml\") pod \"87c34fde-d10d-4c46-a421-80e0ec259618\" (UID: \"87c34fde-d10d-4c46-a421-80e0ec259618\") " Oct 01 09:55:01 crc kubenswrapper[4787]: I1001 09:55:01.646945 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c34fde-d10d-4c46-a421-80e0ec259618-ceilometer-tls-certs\") pod \"87c34fde-d10d-4c46-a421-80e0ec259618\" (UID: \"87c34fde-d10d-4c46-a421-80e0ec259618\") " Oct 01 09:55:01 crc kubenswrapper[4787]: I1001 09:55:01.647050 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87c34fde-d10d-4c46-a421-80e0ec259618-scripts\") pod \"87c34fde-d10d-4c46-a421-80e0ec259618\" (UID: \"87c34fde-d10d-4c46-a421-80e0ec259618\") " Oct 01 09:55:01 crc kubenswrapper[4787]: I1001 09:55:01.647185 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87c34fde-d10d-4c46-a421-80e0ec259618-run-httpd\") pod \"87c34fde-d10d-4c46-a421-80e0ec259618\" (UID: \"87c34fde-d10d-4c46-a421-80e0ec259618\") " Oct 01 09:55:01 crc kubenswrapper[4787]: I1001 09:55:01.647291 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87c34fde-d10d-4c46-a421-80e0ec259618-log-httpd\") pod \"87c34fde-d10d-4c46-a421-80e0ec259618\" (UID: \"87c34fde-d10d-4c46-a421-80e0ec259618\") " Oct 01 09:55:01 crc kubenswrapper[4787]: I1001 09:55:01.647358 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87c34fde-d10d-4c46-a421-80e0ec259618-config-data\") pod \"87c34fde-d10d-4c46-a421-80e0ec259618\" (UID: \"87c34fde-d10d-4c46-a421-80e0ec259618\") " Oct 01 09:55:01 crc kubenswrapper[4787]: I1001 09:55:01.647515 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87c34fde-d10d-4c46-a421-80e0ec259618-combined-ca-bundle\") pod \"87c34fde-d10d-4c46-a421-80e0ec259618\" (UID: \"87c34fde-d10d-4c46-a421-80e0ec259618\") " Oct 01 09:55:01 crc kubenswrapper[4787]: I1001 09:55:01.649069 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87c34fde-d10d-4c46-a421-80e0ec259618-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "87c34fde-d10d-4c46-a421-80e0ec259618" (UID: "87c34fde-d10d-4c46-a421-80e0ec259618"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:55:01 crc kubenswrapper[4787]: I1001 09:55:01.649655 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87c34fde-d10d-4c46-a421-80e0ec259618-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "87c34fde-d10d-4c46-a421-80e0ec259618" (UID: "87c34fde-d10d-4c46-a421-80e0ec259618"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:55:01 crc kubenswrapper[4787]: I1001 09:55:01.658136 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87c34fde-d10d-4c46-a421-80e0ec259618-kube-api-access-q27rn" (OuterVolumeSpecName: "kube-api-access-q27rn") pod "87c34fde-d10d-4c46-a421-80e0ec259618" (UID: "87c34fde-d10d-4c46-a421-80e0ec259618"). InnerVolumeSpecName "kube-api-access-q27rn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:55:01 crc kubenswrapper[4787]: I1001 09:55:01.660889 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87c34fde-d10d-4c46-a421-80e0ec259618-scripts" (OuterVolumeSpecName: "scripts") pod "87c34fde-d10d-4c46-a421-80e0ec259618" (UID: "87c34fde-d10d-4c46-a421-80e0ec259618"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:55:01 crc kubenswrapper[4787]: I1001 09:55:01.693852 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87c34fde-d10d-4c46-a421-80e0ec259618-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "87c34fde-d10d-4c46-a421-80e0ec259618" (UID: "87c34fde-d10d-4c46-a421-80e0ec259618"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:55:01 crc kubenswrapper[4787]: I1001 09:55:01.722594 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87c34fde-d10d-4c46-a421-80e0ec259618-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "87c34fde-d10d-4c46-a421-80e0ec259618" (UID: "87c34fde-d10d-4c46-a421-80e0ec259618"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:55:01 crc kubenswrapper[4787]: I1001 09:55:01.749931 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q27rn\" (UniqueName: \"kubernetes.io/projected/87c34fde-d10d-4c46-a421-80e0ec259618-kube-api-access-q27rn\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:01 crc kubenswrapper[4787]: I1001 09:55:01.749973 4787 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/87c34fde-d10d-4c46-a421-80e0ec259618-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:01 crc kubenswrapper[4787]: I1001 09:55:01.749985 4787 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c34fde-d10d-4c46-a421-80e0ec259618-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:01 crc kubenswrapper[4787]: I1001 09:55:01.749994 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87c34fde-d10d-4c46-a421-80e0ec259618-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:01 crc kubenswrapper[4787]: I1001 09:55:01.750002 4787 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87c34fde-d10d-4c46-a421-80e0ec259618-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:01 crc kubenswrapper[4787]: I1001 09:55:01.750011 4787 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87c34fde-d10d-4c46-a421-80e0ec259618-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:01 crc kubenswrapper[4787]: I1001 09:55:01.750057 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87c34fde-d10d-4c46-a421-80e0ec259618-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "87c34fde-d10d-4c46-a421-80e0ec259618" (UID: "87c34fde-d10d-4c46-a421-80e0ec259618"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:55:01 crc kubenswrapper[4787]: I1001 09:55:01.771976 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87c34fde-d10d-4c46-a421-80e0ec259618-config-data" (OuterVolumeSpecName: "config-data") pod "87c34fde-d10d-4c46-a421-80e0ec259618" (UID: "87c34fde-d10d-4c46-a421-80e0ec259618"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:55:01 crc kubenswrapper[4787]: I1001 09:55:01.851724 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87c34fde-d10d-4c46-a421-80e0ec259618-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:01 crc kubenswrapper[4787]: I1001 09:55:01.851765 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87c34fde-d10d-4c46-a421-80e0ec259618-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.382572 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"87c34fde-d10d-4c46-a421-80e0ec259618","Type":"ContainerDied","Data":"51e61818de9f4b31deb8939edd371f92f07d29cd0984d32557ca9b21406a98ad"} Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.382683 4787 scope.go:117] "RemoveContainer" containerID="9456b2948249c91faffcd699d3f7fec1eea21de1c78841b56f960beba8113270" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.382702 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.422344 4787 scope.go:117] "RemoveContainer" containerID="9fa49d4036bc4bddff9f38920d3c3166c8c166f938fab2d7567edfcd765f035d" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.437689 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.453034 4787 scope.go:117] "RemoveContainer" containerID="2279e2f22be98966a586fefeac39a0f0d83c6396bcfce3feea7247405e00b973" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.458209 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.469389 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:55:02 crc kubenswrapper[4787]: E1001 09:55:02.469908 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87c34fde-d10d-4c46-a421-80e0ec259618" containerName="proxy-httpd" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.469926 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="87c34fde-d10d-4c46-a421-80e0ec259618" containerName="proxy-httpd" Oct 01 09:55:02 crc kubenswrapper[4787]: E1001 09:55:02.469969 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87c34fde-d10d-4c46-a421-80e0ec259618" containerName="sg-core" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.469976 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="87c34fde-d10d-4c46-a421-80e0ec259618" containerName="sg-core" Oct 01 09:55:02 crc kubenswrapper[4787]: E1001 09:55:02.469987 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87c34fde-d10d-4c46-a421-80e0ec259618" containerName="ceilometer-notification-agent" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.469995 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="87c34fde-d10d-4c46-a421-80e0ec259618" containerName="ceilometer-notification-agent" Oct 01 09:55:02 crc kubenswrapper[4787]: E1001 09:55:02.470012 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87c34fde-d10d-4c46-a421-80e0ec259618" containerName="ceilometer-central-agent" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.470019 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="87c34fde-d10d-4c46-a421-80e0ec259618" containerName="ceilometer-central-agent" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.470487 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="87c34fde-d10d-4c46-a421-80e0ec259618" containerName="ceilometer-central-agent" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.470517 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="87c34fde-d10d-4c46-a421-80e0ec259618" containerName="proxy-httpd" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.470527 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="87c34fde-d10d-4c46-a421-80e0ec259618" containerName="sg-core" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.470537 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="87c34fde-d10d-4c46-a421-80e0ec259618" containerName="ceilometer-notification-agent" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.472758 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.473966 4787 scope.go:117] "RemoveContainer" containerID="8d6e18fe98ef9027ff8eb7eb0f6778f4bbc31504c85bea3b6ec35e8b1fda0c99" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.475185 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.476612 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.476765 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.477546 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.538761 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87c34fde-d10d-4c46-a421-80e0ec259618" path="/var/lib/kubelet/pods/87c34fde-d10d-4c46-a421-80e0ec259618/volumes" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.568254 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0cdb08fb-0a31-40dd-98c6-d13926671f45-run-httpd\") pod \"ceilometer-0\" (UID: \"0cdb08fb-0a31-40dd-98c6-d13926671f45\") " pod="openstack/ceilometer-0" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.568307 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0cdb08fb-0a31-40dd-98c6-d13926671f45-scripts\") pod \"ceilometer-0\" (UID: \"0cdb08fb-0a31-40dd-98c6-d13926671f45\") " pod="openstack/ceilometer-0" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.568348 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cdb08fb-0a31-40dd-98c6-d13926671f45-config-data\") pod \"ceilometer-0\" (UID: \"0cdb08fb-0a31-40dd-98c6-d13926671f45\") " pod="openstack/ceilometer-0" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.568601 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cdb08fb-0a31-40dd-98c6-d13926671f45-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0cdb08fb-0a31-40dd-98c6-d13926671f45\") " pod="openstack/ceilometer-0" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.568672 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0cdb08fb-0a31-40dd-98c6-d13926671f45-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0cdb08fb-0a31-40dd-98c6-d13926671f45\") " pod="openstack/ceilometer-0" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.568916 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0cdb08fb-0a31-40dd-98c6-d13926671f45-log-httpd\") pod \"ceilometer-0\" (UID: \"0cdb08fb-0a31-40dd-98c6-d13926671f45\") " pod="openstack/ceilometer-0" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.568996 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0cdb08fb-0a31-40dd-98c6-d13926671f45-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0cdb08fb-0a31-40dd-98c6-d13926671f45\") " pod="openstack/ceilometer-0" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.569014 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rhn8\" (UniqueName: \"kubernetes.io/projected/0cdb08fb-0a31-40dd-98c6-d13926671f45-kube-api-access-7rhn8\") pod \"ceilometer-0\" (UID: \"0cdb08fb-0a31-40dd-98c6-d13926671f45\") " pod="openstack/ceilometer-0" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.671732 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0cdb08fb-0a31-40dd-98c6-d13926671f45-log-httpd\") pod \"ceilometer-0\" (UID: \"0cdb08fb-0a31-40dd-98c6-d13926671f45\") " pod="openstack/ceilometer-0" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.671803 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0cdb08fb-0a31-40dd-98c6-d13926671f45-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0cdb08fb-0a31-40dd-98c6-d13926671f45\") " pod="openstack/ceilometer-0" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.671830 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rhn8\" (UniqueName: \"kubernetes.io/projected/0cdb08fb-0a31-40dd-98c6-d13926671f45-kube-api-access-7rhn8\") pod \"ceilometer-0\" (UID: \"0cdb08fb-0a31-40dd-98c6-d13926671f45\") " pod="openstack/ceilometer-0" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.671868 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0cdb08fb-0a31-40dd-98c6-d13926671f45-run-httpd\") pod \"ceilometer-0\" (UID: \"0cdb08fb-0a31-40dd-98c6-d13926671f45\") " pod="openstack/ceilometer-0" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.671895 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0cdb08fb-0a31-40dd-98c6-d13926671f45-scripts\") pod \"ceilometer-0\" (UID: \"0cdb08fb-0a31-40dd-98c6-d13926671f45\") " pod="openstack/ceilometer-0" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.671923 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cdb08fb-0a31-40dd-98c6-d13926671f45-config-data\") pod \"ceilometer-0\" (UID: \"0cdb08fb-0a31-40dd-98c6-d13926671f45\") " pod="openstack/ceilometer-0" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.671982 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cdb08fb-0a31-40dd-98c6-d13926671f45-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0cdb08fb-0a31-40dd-98c6-d13926671f45\") " pod="openstack/ceilometer-0" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.672011 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0cdb08fb-0a31-40dd-98c6-d13926671f45-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0cdb08fb-0a31-40dd-98c6-d13926671f45\") " pod="openstack/ceilometer-0" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.672409 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0cdb08fb-0a31-40dd-98c6-d13926671f45-run-httpd\") pod \"ceilometer-0\" (UID: \"0cdb08fb-0a31-40dd-98c6-d13926671f45\") " pod="openstack/ceilometer-0" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.674227 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0cdb08fb-0a31-40dd-98c6-d13926671f45-log-httpd\") pod \"ceilometer-0\" (UID: \"0cdb08fb-0a31-40dd-98c6-d13926671f45\") " pod="openstack/ceilometer-0" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.678433 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0cdb08fb-0a31-40dd-98c6-d13926671f45-scripts\") pod \"ceilometer-0\" (UID: \"0cdb08fb-0a31-40dd-98c6-d13926671f45\") " pod="openstack/ceilometer-0" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.679710 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cdb08fb-0a31-40dd-98c6-d13926671f45-config-data\") pod \"ceilometer-0\" (UID: \"0cdb08fb-0a31-40dd-98c6-d13926671f45\") " pod="openstack/ceilometer-0" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.679682 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cdb08fb-0a31-40dd-98c6-d13926671f45-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0cdb08fb-0a31-40dd-98c6-d13926671f45\") " pod="openstack/ceilometer-0" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.687803 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0cdb08fb-0a31-40dd-98c6-d13926671f45-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0cdb08fb-0a31-40dd-98c6-d13926671f45\") " pod="openstack/ceilometer-0" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.688505 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0cdb08fb-0a31-40dd-98c6-d13926671f45-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0cdb08fb-0a31-40dd-98c6-d13926671f45\") " pod="openstack/ceilometer-0" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.692705 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rhn8\" (UniqueName: \"kubernetes.io/projected/0cdb08fb-0a31-40dd-98c6-d13926671f45-kube-api-access-7rhn8\") pod \"ceilometer-0\" (UID: \"0cdb08fb-0a31-40dd-98c6-d13926671f45\") " pod="openstack/ceilometer-0" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.810972 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 01 09:55:02 crc kubenswrapper[4787]: I1001 09:55:02.914902 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.088730 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec425a80-032d-48b0-b46d-a38d98482149-config-data\") pod \"ec425a80-032d-48b0-b46d-a38d98482149\" (UID: \"ec425a80-032d-48b0-b46d-a38d98482149\") " Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.088936 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77hcf\" (UniqueName: \"kubernetes.io/projected/ec425a80-032d-48b0-b46d-a38d98482149-kube-api-access-77hcf\") pod \"ec425a80-032d-48b0-b46d-a38d98482149\" (UID: \"ec425a80-032d-48b0-b46d-a38d98482149\") " Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.088992 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec425a80-032d-48b0-b46d-a38d98482149-logs\") pod \"ec425a80-032d-48b0-b46d-a38d98482149\" (UID: \"ec425a80-032d-48b0-b46d-a38d98482149\") " Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.089043 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec425a80-032d-48b0-b46d-a38d98482149-combined-ca-bundle\") pod \"ec425a80-032d-48b0-b46d-a38d98482149\" (UID: \"ec425a80-032d-48b0-b46d-a38d98482149\") " Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.103563 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec425a80-032d-48b0-b46d-a38d98482149-logs" (OuterVolumeSpecName: "logs") pod "ec425a80-032d-48b0-b46d-a38d98482149" (UID: "ec425a80-032d-48b0-b46d-a38d98482149"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.110316 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec425a80-032d-48b0-b46d-a38d98482149-kube-api-access-77hcf" (OuterVolumeSpecName: "kube-api-access-77hcf") pod "ec425a80-032d-48b0-b46d-a38d98482149" (UID: "ec425a80-032d-48b0-b46d-a38d98482149"). InnerVolumeSpecName "kube-api-access-77hcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.193911 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77hcf\" (UniqueName: \"kubernetes.io/projected/ec425a80-032d-48b0-b46d-a38d98482149-kube-api-access-77hcf\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.194273 4787 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec425a80-032d-48b0-b46d-a38d98482149-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.200324 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec425a80-032d-48b0-b46d-a38d98482149-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ec425a80-032d-48b0-b46d-a38d98482149" (UID: "ec425a80-032d-48b0-b46d-a38d98482149"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.231156 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec425a80-032d-48b0-b46d-a38d98482149-config-data" (OuterVolumeSpecName: "config-data") pod "ec425a80-032d-48b0-b46d-a38d98482149" (UID: "ec425a80-032d-48b0-b46d-a38d98482149"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.297002 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec425a80-032d-48b0-b46d-a38d98482149-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.297040 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec425a80-032d-48b0-b46d-a38d98482149-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.328873 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.394482 4787 generic.go:334] "Generic (PLEG): container finished" podID="ec425a80-032d-48b0-b46d-a38d98482149" containerID="47877f416671132ec65ad4daf6c9681db228d733d74960a9216ae06960eed683" exitCode=0 Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.394587 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.394588 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ec425a80-032d-48b0-b46d-a38d98482149","Type":"ContainerDied","Data":"47877f416671132ec65ad4daf6c9681db228d733d74960a9216ae06960eed683"} Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.394848 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ec425a80-032d-48b0-b46d-a38d98482149","Type":"ContainerDied","Data":"57b2d5c54eff49731d44a6c954a5c971dbb1ac37debcf258b1a81ab5847b4101"} Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.394888 4787 scope.go:117] "RemoveContainer" containerID="47877f416671132ec65ad4daf6c9681db228d733d74960a9216ae06960eed683" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.398621 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0cdb08fb-0a31-40dd-98c6-d13926671f45","Type":"ContainerStarted","Data":"69c52e11a0d9a37c821b4be15acc9d1581780e6c6e8ec28ea81af67b622ff1ea"} Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.438096 4787 scope.go:117] "RemoveContainer" containerID="aac3eccdc8574c073088fc9f79a2eb36d2a4b971b79fead4de1bf5723a6378a6" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.452144 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.467977 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.473979 4787 scope.go:117] "RemoveContainer" containerID="47877f416671132ec65ad4daf6c9681db228d733d74960a9216ae06960eed683" Oct 01 09:55:03 crc kubenswrapper[4787]: E1001 09:55:03.474815 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47877f416671132ec65ad4daf6c9681db228d733d74960a9216ae06960eed683\": container with ID starting with 47877f416671132ec65ad4daf6c9681db228d733d74960a9216ae06960eed683 not found: ID does not exist" containerID="47877f416671132ec65ad4daf6c9681db228d733d74960a9216ae06960eed683" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.474897 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47877f416671132ec65ad4daf6c9681db228d733d74960a9216ae06960eed683"} err="failed to get container status \"47877f416671132ec65ad4daf6c9681db228d733d74960a9216ae06960eed683\": rpc error: code = NotFound desc = could not find container \"47877f416671132ec65ad4daf6c9681db228d733d74960a9216ae06960eed683\": container with ID starting with 47877f416671132ec65ad4daf6c9681db228d733d74960a9216ae06960eed683 not found: ID does not exist" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.474968 4787 scope.go:117] "RemoveContainer" containerID="aac3eccdc8574c073088fc9f79a2eb36d2a4b971b79fead4de1bf5723a6378a6" Oct 01 09:55:03 crc kubenswrapper[4787]: E1001 09:55:03.477575 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aac3eccdc8574c073088fc9f79a2eb36d2a4b971b79fead4de1bf5723a6378a6\": container with ID starting with aac3eccdc8574c073088fc9f79a2eb36d2a4b971b79fead4de1bf5723a6378a6 not found: ID does not exist" containerID="aac3eccdc8574c073088fc9f79a2eb36d2a4b971b79fead4de1bf5723a6378a6" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.477621 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aac3eccdc8574c073088fc9f79a2eb36d2a4b971b79fead4de1bf5723a6378a6"} err="failed to get container status \"aac3eccdc8574c073088fc9f79a2eb36d2a4b971b79fead4de1bf5723a6378a6\": rpc error: code = NotFound desc = could not find container \"aac3eccdc8574c073088fc9f79a2eb36d2a4b971b79fead4de1bf5723a6378a6\": container with ID starting with aac3eccdc8574c073088fc9f79a2eb36d2a4b971b79fead4de1bf5723a6378a6 not found: ID does not exist" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.483240 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 01 09:55:03 crc kubenswrapper[4787]: E1001 09:55:03.483698 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec425a80-032d-48b0-b46d-a38d98482149" containerName="nova-api-log" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.483718 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec425a80-032d-48b0-b46d-a38d98482149" containerName="nova-api-log" Oct 01 09:55:03 crc kubenswrapper[4787]: E1001 09:55:03.483749 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec425a80-032d-48b0-b46d-a38d98482149" containerName="nova-api-api" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.483757 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec425a80-032d-48b0-b46d-a38d98482149" containerName="nova-api-api" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.483930 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec425a80-032d-48b0-b46d-a38d98482149" containerName="nova-api-log" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.483954 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec425a80-032d-48b0-b46d-a38d98482149" containerName="nova-api-api" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.484958 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.488435 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.492211 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.492350 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.492404 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.602014 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/707990a8-7470-4e49-a762-e2d8f5c9382b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"707990a8-7470-4e49-a762-e2d8f5c9382b\") " pod="openstack/nova-api-0" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.602291 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/707990a8-7470-4e49-a762-e2d8f5c9382b-public-tls-certs\") pod \"nova-api-0\" (UID: \"707990a8-7470-4e49-a762-e2d8f5c9382b\") " pod="openstack/nova-api-0" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.602483 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmsxp\" (UniqueName: \"kubernetes.io/projected/707990a8-7470-4e49-a762-e2d8f5c9382b-kube-api-access-fmsxp\") pod \"nova-api-0\" (UID: \"707990a8-7470-4e49-a762-e2d8f5c9382b\") " pod="openstack/nova-api-0" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.602558 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/707990a8-7470-4e49-a762-e2d8f5c9382b-logs\") pod \"nova-api-0\" (UID: \"707990a8-7470-4e49-a762-e2d8f5c9382b\") " pod="openstack/nova-api-0" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.602780 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/707990a8-7470-4e49-a762-e2d8f5c9382b-config-data\") pod \"nova-api-0\" (UID: \"707990a8-7470-4e49-a762-e2d8f5c9382b\") " pod="openstack/nova-api-0" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.602930 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/707990a8-7470-4e49-a762-e2d8f5c9382b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"707990a8-7470-4e49-a762-e2d8f5c9382b\") " pod="openstack/nova-api-0" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.705012 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/707990a8-7470-4e49-a762-e2d8f5c9382b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"707990a8-7470-4e49-a762-e2d8f5c9382b\") " pod="openstack/nova-api-0" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.705123 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/707990a8-7470-4e49-a762-e2d8f5c9382b-public-tls-certs\") pod \"nova-api-0\" (UID: \"707990a8-7470-4e49-a762-e2d8f5c9382b\") " pod="openstack/nova-api-0" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.705193 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmsxp\" (UniqueName: \"kubernetes.io/projected/707990a8-7470-4e49-a762-e2d8f5c9382b-kube-api-access-fmsxp\") pod \"nova-api-0\" (UID: \"707990a8-7470-4e49-a762-e2d8f5c9382b\") " pod="openstack/nova-api-0" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.705229 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/707990a8-7470-4e49-a762-e2d8f5c9382b-logs\") pod \"nova-api-0\" (UID: \"707990a8-7470-4e49-a762-e2d8f5c9382b\") " pod="openstack/nova-api-0" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.705318 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/707990a8-7470-4e49-a762-e2d8f5c9382b-config-data\") pod \"nova-api-0\" (UID: \"707990a8-7470-4e49-a762-e2d8f5c9382b\") " pod="openstack/nova-api-0" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.705372 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/707990a8-7470-4e49-a762-e2d8f5c9382b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"707990a8-7470-4e49-a762-e2d8f5c9382b\") " pod="openstack/nova-api-0" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.706695 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/707990a8-7470-4e49-a762-e2d8f5c9382b-logs\") pod \"nova-api-0\" (UID: \"707990a8-7470-4e49-a762-e2d8f5c9382b\") " pod="openstack/nova-api-0" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.710236 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/707990a8-7470-4e49-a762-e2d8f5c9382b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"707990a8-7470-4e49-a762-e2d8f5c9382b\") " pod="openstack/nova-api-0" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.711117 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/707990a8-7470-4e49-a762-e2d8f5c9382b-config-data\") pod \"nova-api-0\" (UID: \"707990a8-7470-4e49-a762-e2d8f5c9382b\") " pod="openstack/nova-api-0" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.711810 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/707990a8-7470-4e49-a762-e2d8f5c9382b-public-tls-certs\") pod \"nova-api-0\" (UID: \"707990a8-7470-4e49-a762-e2d8f5c9382b\") " pod="openstack/nova-api-0" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.722906 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/707990a8-7470-4e49-a762-e2d8f5c9382b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"707990a8-7470-4e49-a762-e2d8f5c9382b\") " pod="openstack/nova-api-0" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.730743 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmsxp\" (UniqueName: \"kubernetes.io/projected/707990a8-7470-4e49-a762-e2d8f5c9382b-kube-api-access-fmsxp\") pod \"nova-api-0\" (UID: \"707990a8-7470-4e49-a762-e2d8f5c9382b\") " pod="openstack/nova-api-0" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.804719 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.851029 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:55:03 crc kubenswrapper[4787]: I1001 09:55:03.876723 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:55:04 crc kubenswrapper[4787]: I1001 09:55:04.306339 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 09:55:04 crc kubenswrapper[4787]: W1001 09:55:04.310956 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod707990a8_7470_4e49_a762_e2d8f5c9382b.slice/crio-ec28c418739979c0a5de9ad66c8f79662f50c33e2980d59599126a4742fae79d WatchSource:0}: Error finding container ec28c418739979c0a5de9ad66c8f79662f50c33e2980d59599126a4742fae79d: Status 404 returned error can't find the container with id ec28c418739979c0a5de9ad66c8f79662f50c33e2980d59599126a4742fae79d Oct 01 09:55:04 crc kubenswrapper[4787]: I1001 09:55:04.431489 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"707990a8-7470-4e49-a762-e2d8f5c9382b","Type":"ContainerStarted","Data":"ec28c418739979c0a5de9ad66c8f79662f50c33e2980d59599126a4742fae79d"} Oct 01 09:55:04 crc kubenswrapper[4787]: I1001 09:55:04.437149 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0cdb08fb-0a31-40dd-98c6-d13926671f45","Type":"ContainerStarted","Data":"0df9f85b0d9029b3baec05720d6fec0c57b98ddbbaefe5f6040df8cf23549601"} Oct 01 09:55:04 crc kubenswrapper[4787]: I1001 09:55:04.453207 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 01 09:55:04 crc kubenswrapper[4787]: I1001 09:55:04.538647 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec425a80-032d-48b0-b46d-a38d98482149" path="/var/lib/kubelet/pods/ec425a80-032d-48b0-b46d-a38d98482149/volumes" Oct 01 09:55:04 crc kubenswrapper[4787]: I1001 09:55:04.670444 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 01 09:55:04 crc kubenswrapper[4787]: I1001 09:55:04.670930 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 01 09:55:04 crc kubenswrapper[4787]: I1001 09:55:04.737222 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-8655l"] Oct 01 09:55:04 crc kubenswrapper[4787]: I1001 09:55:04.739053 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-8655l" Oct 01 09:55:04 crc kubenswrapper[4787]: I1001 09:55:04.742965 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-8655l"] Oct 01 09:55:04 crc kubenswrapper[4787]: I1001 09:55:04.771783 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 01 09:55:04 crc kubenswrapper[4787]: I1001 09:55:04.771966 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 01 09:55:04 crc kubenswrapper[4787]: I1001 09:55:04.844141 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5d060f6-9fac-4c41-b365-033b09e68016-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-8655l\" (UID: \"f5d060f6-9fac-4c41-b365-033b09e68016\") " pod="openstack/nova-cell1-cell-mapping-8655l" Oct 01 09:55:04 crc kubenswrapper[4787]: I1001 09:55:04.844229 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5d060f6-9fac-4c41-b365-033b09e68016-config-data\") pod \"nova-cell1-cell-mapping-8655l\" (UID: \"f5d060f6-9fac-4c41-b365-033b09e68016\") " pod="openstack/nova-cell1-cell-mapping-8655l" Oct 01 09:55:04 crc kubenswrapper[4787]: I1001 09:55:04.844412 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txggt\" (UniqueName: \"kubernetes.io/projected/f5d060f6-9fac-4c41-b365-033b09e68016-kube-api-access-txggt\") pod \"nova-cell1-cell-mapping-8655l\" (UID: \"f5d060f6-9fac-4c41-b365-033b09e68016\") " pod="openstack/nova-cell1-cell-mapping-8655l" Oct 01 09:55:04 crc kubenswrapper[4787]: I1001 09:55:04.844780 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5d060f6-9fac-4c41-b365-033b09e68016-scripts\") pod \"nova-cell1-cell-mapping-8655l\" (UID: \"f5d060f6-9fac-4c41-b365-033b09e68016\") " pod="openstack/nova-cell1-cell-mapping-8655l" Oct 01 09:55:04 crc kubenswrapper[4787]: I1001 09:55:04.947036 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5d060f6-9fac-4c41-b365-033b09e68016-scripts\") pod \"nova-cell1-cell-mapping-8655l\" (UID: \"f5d060f6-9fac-4c41-b365-033b09e68016\") " pod="openstack/nova-cell1-cell-mapping-8655l" Oct 01 09:55:04 crc kubenswrapper[4787]: I1001 09:55:04.947151 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5d060f6-9fac-4c41-b365-033b09e68016-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-8655l\" (UID: \"f5d060f6-9fac-4c41-b365-033b09e68016\") " pod="openstack/nova-cell1-cell-mapping-8655l" Oct 01 09:55:04 crc kubenswrapper[4787]: I1001 09:55:04.947194 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5d060f6-9fac-4c41-b365-033b09e68016-config-data\") pod \"nova-cell1-cell-mapping-8655l\" (UID: \"f5d060f6-9fac-4c41-b365-033b09e68016\") " pod="openstack/nova-cell1-cell-mapping-8655l" Oct 01 09:55:04 crc kubenswrapper[4787]: I1001 09:55:04.947253 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txggt\" (UniqueName: \"kubernetes.io/projected/f5d060f6-9fac-4c41-b365-033b09e68016-kube-api-access-txggt\") pod \"nova-cell1-cell-mapping-8655l\" (UID: \"f5d060f6-9fac-4c41-b365-033b09e68016\") " pod="openstack/nova-cell1-cell-mapping-8655l" Oct 01 09:55:04 crc kubenswrapper[4787]: I1001 09:55:04.951784 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5d060f6-9fac-4c41-b365-033b09e68016-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-8655l\" (UID: \"f5d060f6-9fac-4c41-b365-033b09e68016\") " pod="openstack/nova-cell1-cell-mapping-8655l" Oct 01 09:55:04 crc kubenswrapper[4787]: I1001 09:55:04.953228 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5d060f6-9fac-4c41-b365-033b09e68016-config-data\") pod \"nova-cell1-cell-mapping-8655l\" (UID: \"f5d060f6-9fac-4c41-b365-033b09e68016\") " pod="openstack/nova-cell1-cell-mapping-8655l" Oct 01 09:55:04 crc kubenswrapper[4787]: I1001 09:55:04.953867 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5d060f6-9fac-4c41-b365-033b09e68016-scripts\") pod \"nova-cell1-cell-mapping-8655l\" (UID: \"f5d060f6-9fac-4c41-b365-033b09e68016\") " pod="openstack/nova-cell1-cell-mapping-8655l" Oct 01 09:55:04 crc kubenswrapper[4787]: I1001 09:55:04.966542 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txggt\" (UniqueName: \"kubernetes.io/projected/f5d060f6-9fac-4c41-b365-033b09e68016-kube-api-access-txggt\") pod \"nova-cell1-cell-mapping-8655l\" (UID: \"f5d060f6-9fac-4c41-b365-033b09e68016\") " pod="openstack/nova-cell1-cell-mapping-8655l" Oct 01 09:55:05 crc kubenswrapper[4787]: I1001 09:55:05.101351 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-8655l" Oct 01 09:55:05 crc kubenswrapper[4787]: I1001 09:55:05.447438 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0cdb08fb-0a31-40dd-98c6-d13926671f45","Type":"ContainerStarted","Data":"7a177c932ed32e738f8e26be6b772e5c5bd6a85452d871af6f0ce07788c0ca3c"} Oct 01 09:55:05 crc kubenswrapper[4787]: I1001 09:55:05.456241 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"707990a8-7470-4e49-a762-e2d8f5c9382b","Type":"ContainerStarted","Data":"a3fcb4555ea22b725429b6f79627e1870ca115f8c8df39c56ec3c1285cb0fbc8"} Oct 01 09:55:05 crc kubenswrapper[4787]: I1001 09:55:05.456299 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"707990a8-7470-4e49-a762-e2d8f5c9382b","Type":"ContainerStarted","Data":"14862f4ce85878f8cca6c4b4f5ac3bb1723e878ef53a6bbf9cccc355fb909d47"} Oct 01 09:55:05 crc kubenswrapper[4787]: I1001 09:55:05.496213 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.496176945 podStartE2EDuration="2.496176945s" podCreationTimestamp="2025-10-01 09:55:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:55:05.491979218 +0000 UTC m=+1137.607123395" watchObservedRunningTime="2025-10-01 09:55:05.496176945 +0000 UTC m=+1137.611321102" Oct 01 09:55:05 crc kubenswrapper[4787]: I1001 09:55:05.610818 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-8655l"] Oct 01 09:55:05 crc kubenswrapper[4787]: I1001 09:55:05.688287 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="700e3732-7195-48fb-a5f6-4fb80b9f13e8" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 09:55:05 crc kubenswrapper[4787]: I1001 09:55:05.688910 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="700e3732-7195-48fb-a5f6-4fb80b9f13e8" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 09:55:06 crc kubenswrapper[4787]: I1001 09:55:06.468390 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0cdb08fb-0a31-40dd-98c6-d13926671f45","Type":"ContainerStarted","Data":"43579a5e855fccae6381b91afbf83872cf543e0b35a610362697e8523bea3729"} Oct 01 09:55:06 crc kubenswrapper[4787]: I1001 09:55:06.470744 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-8655l" event={"ID":"f5d060f6-9fac-4c41-b365-033b09e68016","Type":"ContainerStarted","Data":"9d52221a28c0b16ca76146635754207f0869a14d5e5187ba7d4afc4661e743a4"} Oct 01 09:55:06 crc kubenswrapper[4787]: I1001 09:55:06.470796 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-8655l" event={"ID":"f5d060f6-9fac-4c41-b365-033b09e68016","Type":"ContainerStarted","Data":"a7a17f73afcfc96430e13a8e6b2b27d457b167a39457ba40ed5b46ee90c19168"} Oct 01 09:55:06 crc kubenswrapper[4787]: I1001 09:55:06.489622 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-8655l" podStartSLOduration=2.489603048 podStartE2EDuration="2.489603048s" podCreationTimestamp="2025-10-01 09:55:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:55:06.4869215 +0000 UTC m=+1138.602065687" watchObservedRunningTime="2025-10-01 09:55:06.489603048 +0000 UTC m=+1138.604747205" Oct 01 09:55:06 crc kubenswrapper[4787]: I1001 09:55:06.868336 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-68b59d98cf-xk49b" Oct 01 09:55:06 crc kubenswrapper[4787]: I1001 09:55:06.966301 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d654f6767-wtpqm"] Oct 01 09:55:06 crc kubenswrapper[4787]: I1001 09:55:06.967107 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5d654f6767-wtpqm" podUID="e9574c0c-29bf-43bc-b090-35ece3a54e94" containerName="dnsmasq-dns" containerID="cri-o://2e0b5dd5073c0fa32a8d6df15f44078a84e399426c318ffd3e06018034b44a89" gracePeriod=10 Oct 01 09:55:07 crc kubenswrapper[4787]: I1001 09:55:07.493221 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0cdb08fb-0a31-40dd-98c6-d13926671f45","Type":"ContainerStarted","Data":"8da351b0f0bf7e9a2c89d7c8e4419d140f576ca027a8a342d6d7e2c265f9931b"} Oct 01 09:55:07 crc kubenswrapper[4787]: I1001 09:55:07.495005 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 01 09:55:07 crc kubenswrapper[4787]: I1001 09:55:07.498154 4787 generic.go:334] "Generic (PLEG): container finished" podID="e9574c0c-29bf-43bc-b090-35ece3a54e94" containerID="2e0b5dd5073c0fa32a8d6df15f44078a84e399426c318ffd3e06018034b44a89" exitCode=0 Oct 01 09:55:07 crc kubenswrapper[4787]: I1001 09:55:07.498626 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d654f6767-wtpqm" event={"ID":"e9574c0c-29bf-43bc-b090-35ece3a54e94","Type":"ContainerDied","Data":"2e0b5dd5073c0fa32a8d6df15f44078a84e399426c318ffd3e06018034b44a89"} Oct 01 09:55:07 crc kubenswrapper[4787]: I1001 09:55:07.498772 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d654f6767-wtpqm" event={"ID":"e9574c0c-29bf-43bc-b090-35ece3a54e94","Type":"ContainerDied","Data":"110b081b814c531a86a8e3a9b5b7bc79dde6b5ced503abe1f979ebf41462ebdd"} Oct 01 09:55:07 crc kubenswrapper[4787]: I1001 09:55:07.498790 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="110b081b814c531a86a8e3a9b5b7bc79dde6b5ced503abe1f979ebf41462ebdd" Oct 01 09:55:07 crc kubenswrapper[4787]: I1001 09:55:07.526052 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.880659994 podStartE2EDuration="5.526028422s" podCreationTimestamp="2025-10-01 09:55:02 +0000 UTC" firstStartedPulling="2025-10-01 09:55:03.33755477 +0000 UTC m=+1135.452698917" lastFinishedPulling="2025-10-01 09:55:06.982923178 +0000 UTC m=+1139.098067345" observedRunningTime="2025-10-01 09:55:07.519363413 +0000 UTC m=+1139.634507600" watchObservedRunningTime="2025-10-01 09:55:07.526028422 +0000 UTC m=+1139.641172579" Oct 01 09:55:07 crc kubenswrapper[4787]: I1001 09:55:07.552773 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d654f6767-wtpqm" Oct 01 09:55:07 crc kubenswrapper[4787]: I1001 09:55:07.626556 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e9574c0c-29bf-43bc-b090-35ece3a54e94-dns-swift-storage-0\") pod \"e9574c0c-29bf-43bc-b090-35ece3a54e94\" (UID: \"e9574c0c-29bf-43bc-b090-35ece3a54e94\") " Oct 01 09:55:07 crc kubenswrapper[4787]: I1001 09:55:07.626710 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9574c0c-29bf-43bc-b090-35ece3a54e94-config\") pod \"e9574c0c-29bf-43bc-b090-35ece3a54e94\" (UID: \"e9574c0c-29bf-43bc-b090-35ece3a54e94\") " Oct 01 09:55:07 crc kubenswrapper[4787]: I1001 09:55:07.626778 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hpmqh\" (UniqueName: \"kubernetes.io/projected/e9574c0c-29bf-43bc-b090-35ece3a54e94-kube-api-access-hpmqh\") pod \"e9574c0c-29bf-43bc-b090-35ece3a54e94\" (UID: \"e9574c0c-29bf-43bc-b090-35ece3a54e94\") " Oct 01 09:55:07 crc kubenswrapper[4787]: I1001 09:55:07.626859 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e9574c0c-29bf-43bc-b090-35ece3a54e94-ovsdbserver-nb\") pod \"e9574c0c-29bf-43bc-b090-35ece3a54e94\" (UID: \"e9574c0c-29bf-43bc-b090-35ece3a54e94\") " Oct 01 09:55:07 crc kubenswrapper[4787]: I1001 09:55:07.626999 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e9574c0c-29bf-43bc-b090-35ece3a54e94-ovsdbserver-sb\") pod \"e9574c0c-29bf-43bc-b090-35ece3a54e94\" (UID: \"e9574c0c-29bf-43bc-b090-35ece3a54e94\") " Oct 01 09:55:07 crc kubenswrapper[4787]: I1001 09:55:07.627026 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e9574c0c-29bf-43bc-b090-35ece3a54e94-dns-svc\") pod \"e9574c0c-29bf-43bc-b090-35ece3a54e94\" (UID: \"e9574c0c-29bf-43bc-b090-35ece3a54e94\") " Oct 01 09:55:07 crc kubenswrapper[4787]: I1001 09:55:07.634759 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9574c0c-29bf-43bc-b090-35ece3a54e94-kube-api-access-hpmqh" (OuterVolumeSpecName: "kube-api-access-hpmqh") pod "e9574c0c-29bf-43bc-b090-35ece3a54e94" (UID: "e9574c0c-29bf-43bc-b090-35ece3a54e94"). InnerVolumeSpecName "kube-api-access-hpmqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:55:07 crc kubenswrapper[4787]: I1001 09:55:07.679386 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9574c0c-29bf-43bc-b090-35ece3a54e94-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e9574c0c-29bf-43bc-b090-35ece3a54e94" (UID: "e9574c0c-29bf-43bc-b090-35ece3a54e94"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:55:07 crc kubenswrapper[4787]: I1001 09:55:07.683640 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9574c0c-29bf-43bc-b090-35ece3a54e94-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e9574c0c-29bf-43bc-b090-35ece3a54e94" (UID: "e9574c0c-29bf-43bc-b090-35ece3a54e94"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:55:07 crc kubenswrapper[4787]: I1001 09:55:07.685603 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9574c0c-29bf-43bc-b090-35ece3a54e94-config" (OuterVolumeSpecName: "config") pod "e9574c0c-29bf-43bc-b090-35ece3a54e94" (UID: "e9574c0c-29bf-43bc-b090-35ece3a54e94"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:55:07 crc kubenswrapper[4787]: I1001 09:55:07.699382 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9574c0c-29bf-43bc-b090-35ece3a54e94-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e9574c0c-29bf-43bc-b090-35ece3a54e94" (UID: "e9574c0c-29bf-43bc-b090-35ece3a54e94"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:55:07 crc kubenswrapper[4787]: I1001 09:55:07.719590 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9574c0c-29bf-43bc-b090-35ece3a54e94-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e9574c0c-29bf-43bc-b090-35ece3a54e94" (UID: "e9574c0c-29bf-43bc-b090-35ece3a54e94"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:55:07 crc kubenswrapper[4787]: I1001 09:55:07.730484 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e9574c0c-29bf-43bc-b090-35ece3a54e94-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:07 crc kubenswrapper[4787]: I1001 09:55:07.730537 4787 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e9574c0c-29bf-43bc-b090-35ece3a54e94-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:07 crc kubenswrapper[4787]: I1001 09:55:07.730562 4787 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e9574c0c-29bf-43bc-b090-35ece3a54e94-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:07 crc kubenswrapper[4787]: I1001 09:55:07.730577 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9574c0c-29bf-43bc-b090-35ece3a54e94-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:07 crc kubenswrapper[4787]: I1001 09:55:07.730590 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hpmqh\" (UniqueName: \"kubernetes.io/projected/e9574c0c-29bf-43bc-b090-35ece3a54e94-kube-api-access-hpmqh\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:07 crc kubenswrapper[4787]: I1001 09:55:07.730604 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e9574c0c-29bf-43bc-b090-35ece3a54e94-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:08 crc kubenswrapper[4787]: I1001 09:55:08.511787 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d654f6767-wtpqm" Oct 01 09:55:08 crc kubenswrapper[4787]: I1001 09:55:08.581313 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d654f6767-wtpqm"] Oct 01 09:55:08 crc kubenswrapper[4787]: I1001 09:55:08.592459 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5d654f6767-wtpqm"] Oct 01 09:55:10 crc kubenswrapper[4787]: I1001 09:55:10.556995 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9574c0c-29bf-43bc-b090-35ece3a54e94" path="/var/lib/kubelet/pods/e9574c0c-29bf-43bc-b090-35ece3a54e94/volumes" Oct 01 09:55:11 crc kubenswrapper[4787]: I1001 09:55:11.550038 4787 generic.go:334] "Generic (PLEG): container finished" podID="f5d060f6-9fac-4c41-b365-033b09e68016" containerID="9d52221a28c0b16ca76146635754207f0869a14d5e5187ba7d4afc4661e743a4" exitCode=0 Oct 01 09:55:11 crc kubenswrapper[4787]: I1001 09:55:11.550141 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-8655l" event={"ID":"f5d060f6-9fac-4c41-b365-033b09e68016","Type":"ContainerDied","Data":"9d52221a28c0b16ca76146635754207f0869a14d5e5187ba7d4afc4661e743a4"} Oct 01 09:55:12 crc kubenswrapper[4787]: I1001 09:55:12.496232 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5d654f6767-wtpqm" podUID="e9574c0c-29bf-43bc-b090-35ece3a54e94" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.192:5353: i/o timeout" Oct 01 09:55:12 crc kubenswrapper[4787]: I1001 09:55:12.923433 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-8655l" Oct 01 09:55:13 crc kubenswrapper[4787]: I1001 09:55:13.060101 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5d060f6-9fac-4c41-b365-033b09e68016-combined-ca-bundle\") pod \"f5d060f6-9fac-4c41-b365-033b09e68016\" (UID: \"f5d060f6-9fac-4c41-b365-033b09e68016\") " Oct 01 09:55:13 crc kubenswrapper[4787]: I1001 09:55:13.060185 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5d060f6-9fac-4c41-b365-033b09e68016-scripts\") pod \"f5d060f6-9fac-4c41-b365-033b09e68016\" (UID: \"f5d060f6-9fac-4c41-b365-033b09e68016\") " Oct 01 09:55:13 crc kubenswrapper[4787]: I1001 09:55:13.060225 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5d060f6-9fac-4c41-b365-033b09e68016-config-data\") pod \"f5d060f6-9fac-4c41-b365-033b09e68016\" (UID: \"f5d060f6-9fac-4c41-b365-033b09e68016\") " Oct 01 09:55:13 crc kubenswrapper[4787]: I1001 09:55:13.060356 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txggt\" (UniqueName: \"kubernetes.io/projected/f5d060f6-9fac-4c41-b365-033b09e68016-kube-api-access-txggt\") pod \"f5d060f6-9fac-4c41-b365-033b09e68016\" (UID: \"f5d060f6-9fac-4c41-b365-033b09e68016\") " Oct 01 09:55:13 crc kubenswrapper[4787]: I1001 09:55:13.070626 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5d060f6-9fac-4c41-b365-033b09e68016-scripts" (OuterVolumeSpecName: "scripts") pod "f5d060f6-9fac-4c41-b365-033b09e68016" (UID: "f5d060f6-9fac-4c41-b365-033b09e68016"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:55:13 crc kubenswrapper[4787]: I1001 09:55:13.071345 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5d060f6-9fac-4c41-b365-033b09e68016-kube-api-access-txggt" (OuterVolumeSpecName: "kube-api-access-txggt") pod "f5d060f6-9fac-4c41-b365-033b09e68016" (UID: "f5d060f6-9fac-4c41-b365-033b09e68016"). InnerVolumeSpecName "kube-api-access-txggt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:55:13 crc kubenswrapper[4787]: I1001 09:55:13.088728 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5d060f6-9fac-4c41-b365-033b09e68016-config-data" (OuterVolumeSpecName: "config-data") pod "f5d060f6-9fac-4c41-b365-033b09e68016" (UID: "f5d060f6-9fac-4c41-b365-033b09e68016"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:55:13 crc kubenswrapper[4787]: I1001 09:55:13.094198 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5d060f6-9fac-4c41-b365-033b09e68016-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f5d060f6-9fac-4c41-b365-033b09e68016" (UID: "f5d060f6-9fac-4c41-b365-033b09e68016"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:55:13 crc kubenswrapper[4787]: I1001 09:55:13.163028 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txggt\" (UniqueName: \"kubernetes.io/projected/f5d060f6-9fac-4c41-b365-033b09e68016-kube-api-access-txggt\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:13 crc kubenswrapper[4787]: I1001 09:55:13.163065 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5d060f6-9fac-4c41-b365-033b09e68016-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:13 crc kubenswrapper[4787]: I1001 09:55:13.163089 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5d060f6-9fac-4c41-b365-033b09e68016-scripts\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:13 crc kubenswrapper[4787]: I1001 09:55:13.163098 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5d060f6-9fac-4c41-b365-033b09e68016-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:13 crc kubenswrapper[4787]: I1001 09:55:13.571140 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-8655l" event={"ID":"f5d060f6-9fac-4c41-b365-033b09e68016","Type":"ContainerDied","Data":"a7a17f73afcfc96430e13a8e6b2b27d457b167a39457ba40ed5b46ee90c19168"} Oct 01 09:55:13 crc kubenswrapper[4787]: I1001 09:55:13.571195 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7a17f73afcfc96430e13a8e6b2b27d457b167a39457ba40ed5b46ee90c19168" Oct 01 09:55:13 crc kubenswrapper[4787]: I1001 09:55:13.571238 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-8655l" Oct 01 09:55:13 crc kubenswrapper[4787]: I1001 09:55:13.697026 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 01 09:55:13 crc kubenswrapper[4787]: I1001 09:55:13.697522 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="707990a8-7470-4e49-a762-e2d8f5c9382b" containerName="nova-api-log" containerID="cri-o://14862f4ce85878f8cca6c4b4f5ac3bb1723e878ef53a6bbf9cccc355fb909d47" gracePeriod=30 Oct 01 09:55:13 crc kubenswrapper[4787]: I1001 09:55:13.697602 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="707990a8-7470-4e49-a762-e2d8f5c9382b" containerName="nova-api-api" containerID="cri-o://a3fcb4555ea22b725429b6f79627e1870ca115f8c8df39c56ec3c1285cb0fbc8" gracePeriod=30 Oct 01 09:55:13 crc kubenswrapper[4787]: I1001 09:55:13.716529 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 09:55:13 crc kubenswrapper[4787]: I1001 09:55:13.716989 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="c15f26a3-6250-43c9-822a-974a2f3c0a63" containerName="nova-scheduler-scheduler" containerID="cri-o://e1033e70a9ecb6214b67b36775d24b8575db35d5df073e0b8ff0126db0442282" gracePeriod=30 Oct 01 09:55:13 crc kubenswrapper[4787]: I1001 09:55:13.727522 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:55:13 crc kubenswrapper[4787]: I1001 09:55:13.727826 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="700e3732-7195-48fb-a5f6-4fb80b9f13e8" containerName="nova-metadata-log" containerID="cri-o://5a17511d23a3c2c840e42020b72de1cd5076c84cd25e857681d078bc8332adca" gracePeriod=30 Oct 01 09:55:13 crc kubenswrapper[4787]: I1001 09:55:13.727997 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="700e3732-7195-48fb-a5f6-4fb80b9f13e8" containerName="nova-metadata-metadata" containerID="cri-o://9a4e6b2ebf2a9aee84a3e28a27294b0ceba00c22a18adce42fe141eb75395696" gracePeriod=30 Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.474548 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.583886 4787 generic.go:334] "Generic (PLEG): container finished" podID="c15f26a3-6250-43c9-822a-974a2f3c0a63" containerID="e1033e70a9ecb6214b67b36775d24b8575db35d5df073e0b8ff0126db0442282" exitCode=0 Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.583942 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c15f26a3-6250-43c9-822a-974a2f3c0a63","Type":"ContainerDied","Data":"e1033e70a9ecb6214b67b36775d24b8575db35d5df073e0b8ff0126db0442282"} Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.585892 4787 generic.go:334] "Generic (PLEG): container finished" podID="700e3732-7195-48fb-a5f6-4fb80b9f13e8" containerID="5a17511d23a3c2c840e42020b72de1cd5076c84cd25e857681d078bc8332adca" exitCode=143 Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.585980 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"700e3732-7195-48fb-a5f6-4fb80b9f13e8","Type":"ContainerDied","Data":"5a17511d23a3c2c840e42020b72de1cd5076c84cd25e857681d078bc8332adca"} Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.587557 4787 generic.go:334] "Generic (PLEG): container finished" podID="707990a8-7470-4e49-a762-e2d8f5c9382b" containerID="a3fcb4555ea22b725429b6f79627e1870ca115f8c8df39c56ec3c1285cb0fbc8" exitCode=0 Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.587583 4787 generic.go:334] "Generic (PLEG): container finished" podID="707990a8-7470-4e49-a762-e2d8f5c9382b" containerID="14862f4ce85878f8cca6c4b4f5ac3bb1723e878ef53a6bbf9cccc355fb909d47" exitCode=143 Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.587601 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"707990a8-7470-4e49-a762-e2d8f5c9382b","Type":"ContainerDied","Data":"a3fcb4555ea22b725429b6f79627e1870ca115f8c8df39c56ec3c1285cb0fbc8"} Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.587618 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"707990a8-7470-4e49-a762-e2d8f5c9382b","Type":"ContainerDied","Data":"14862f4ce85878f8cca6c4b4f5ac3bb1723e878ef53a6bbf9cccc355fb909d47"} Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.587629 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"707990a8-7470-4e49-a762-e2d8f5c9382b","Type":"ContainerDied","Data":"ec28c418739979c0a5de9ad66c8f79662f50c33e2980d59599126a4742fae79d"} Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.587648 4787 scope.go:117] "RemoveContainer" containerID="a3fcb4555ea22b725429b6f79627e1870ca115f8c8df39c56ec3c1285cb0fbc8" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.587838 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.590836 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/707990a8-7470-4e49-a762-e2d8f5c9382b-config-data\") pod \"707990a8-7470-4e49-a762-e2d8f5c9382b\" (UID: \"707990a8-7470-4e49-a762-e2d8f5c9382b\") " Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.591877 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmsxp\" (UniqueName: \"kubernetes.io/projected/707990a8-7470-4e49-a762-e2d8f5c9382b-kube-api-access-fmsxp\") pod \"707990a8-7470-4e49-a762-e2d8f5c9382b\" (UID: \"707990a8-7470-4e49-a762-e2d8f5c9382b\") " Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.592113 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/707990a8-7470-4e49-a762-e2d8f5c9382b-combined-ca-bundle\") pod \"707990a8-7470-4e49-a762-e2d8f5c9382b\" (UID: \"707990a8-7470-4e49-a762-e2d8f5c9382b\") " Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.592196 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/707990a8-7470-4e49-a762-e2d8f5c9382b-internal-tls-certs\") pod \"707990a8-7470-4e49-a762-e2d8f5c9382b\" (UID: \"707990a8-7470-4e49-a762-e2d8f5c9382b\") " Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.592344 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/707990a8-7470-4e49-a762-e2d8f5c9382b-public-tls-certs\") pod \"707990a8-7470-4e49-a762-e2d8f5c9382b\" (UID: \"707990a8-7470-4e49-a762-e2d8f5c9382b\") " Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.592388 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/707990a8-7470-4e49-a762-e2d8f5c9382b-logs\") pod \"707990a8-7470-4e49-a762-e2d8f5c9382b\" (UID: \"707990a8-7470-4e49-a762-e2d8f5c9382b\") " Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.597857 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/707990a8-7470-4e49-a762-e2d8f5c9382b-kube-api-access-fmsxp" (OuterVolumeSpecName: "kube-api-access-fmsxp") pod "707990a8-7470-4e49-a762-e2d8f5c9382b" (UID: "707990a8-7470-4e49-a762-e2d8f5c9382b"). InnerVolumeSpecName "kube-api-access-fmsxp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.607507 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/707990a8-7470-4e49-a762-e2d8f5c9382b-logs" (OuterVolumeSpecName: "logs") pod "707990a8-7470-4e49-a762-e2d8f5c9382b" (UID: "707990a8-7470-4e49-a762-e2d8f5c9382b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.624399 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/707990a8-7470-4e49-a762-e2d8f5c9382b-config-data" (OuterVolumeSpecName: "config-data") pod "707990a8-7470-4e49-a762-e2d8f5c9382b" (UID: "707990a8-7470-4e49-a762-e2d8f5c9382b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.627331 4787 scope.go:117] "RemoveContainer" containerID="14862f4ce85878f8cca6c4b4f5ac3bb1723e878ef53a6bbf9cccc355fb909d47" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.642377 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/707990a8-7470-4e49-a762-e2d8f5c9382b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "707990a8-7470-4e49-a762-e2d8f5c9382b" (UID: "707990a8-7470-4e49-a762-e2d8f5c9382b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.642743 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.652766 4787 scope.go:117] "RemoveContainer" containerID="a3fcb4555ea22b725429b6f79627e1870ca115f8c8df39c56ec3c1285cb0fbc8" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.652841 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/707990a8-7470-4e49-a762-e2d8f5c9382b-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "707990a8-7470-4e49-a762-e2d8f5c9382b" (UID: "707990a8-7470-4e49-a762-e2d8f5c9382b"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:55:14 crc kubenswrapper[4787]: E1001 09:55:14.653448 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3fcb4555ea22b725429b6f79627e1870ca115f8c8df39c56ec3c1285cb0fbc8\": container with ID starting with a3fcb4555ea22b725429b6f79627e1870ca115f8c8df39c56ec3c1285cb0fbc8 not found: ID does not exist" containerID="a3fcb4555ea22b725429b6f79627e1870ca115f8c8df39c56ec3c1285cb0fbc8" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.653499 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3fcb4555ea22b725429b6f79627e1870ca115f8c8df39c56ec3c1285cb0fbc8"} err="failed to get container status \"a3fcb4555ea22b725429b6f79627e1870ca115f8c8df39c56ec3c1285cb0fbc8\": rpc error: code = NotFound desc = could not find container \"a3fcb4555ea22b725429b6f79627e1870ca115f8c8df39c56ec3c1285cb0fbc8\": container with ID starting with a3fcb4555ea22b725429b6f79627e1870ca115f8c8df39c56ec3c1285cb0fbc8 not found: ID does not exist" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.653530 4787 scope.go:117] "RemoveContainer" containerID="14862f4ce85878f8cca6c4b4f5ac3bb1723e878ef53a6bbf9cccc355fb909d47" Oct 01 09:55:14 crc kubenswrapper[4787]: E1001 09:55:14.653889 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14862f4ce85878f8cca6c4b4f5ac3bb1723e878ef53a6bbf9cccc355fb909d47\": container with ID starting with 14862f4ce85878f8cca6c4b4f5ac3bb1723e878ef53a6bbf9cccc355fb909d47 not found: ID does not exist" containerID="14862f4ce85878f8cca6c4b4f5ac3bb1723e878ef53a6bbf9cccc355fb909d47" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.653930 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14862f4ce85878f8cca6c4b4f5ac3bb1723e878ef53a6bbf9cccc355fb909d47"} err="failed to get container status \"14862f4ce85878f8cca6c4b4f5ac3bb1723e878ef53a6bbf9cccc355fb909d47\": rpc error: code = NotFound desc = could not find container \"14862f4ce85878f8cca6c4b4f5ac3bb1723e878ef53a6bbf9cccc355fb909d47\": container with ID starting with 14862f4ce85878f8cca6c4b4f5ac3bb1723e878ef53a6bbf9cccc355fb909d47 not found: ID does not exist" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.653959 4787 scope.go:117] "RemoveContainer" containerID="a3fcb4555ea22b725429b6f79627e1870ca115f8c8df39c56ec3c1285cb0fbc8" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.655109 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3fcb4555ea22b725429b6f79627e1870ca115f8c8df39c56ec3c1285cb0fbc8"} err="failed to get container status \"a3fcb4555ea22b725429b6f79627e1870ca115f8c8df39c56ec3c1285cb0fbc8\": rpc error: code = NotFound desc = could not find container \"a3fcb4555ea22b725429b6f79627e1870ca115f8c8df39c56ec3c1285cb0fbc8\": container with ID starting with a3fcb4555ea22b725429b6f79627e1870ca115f8c8df39c56ec3c1285cb0fbc8 not found: ID does not exist" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.655136 4787 scope.go:117] "RemoveContainer" containerID="14862f4ce85878f8cca6c4b4f5ac3bb1723e878ef53a6bbf9cccc355fb909d47" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.656790 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14862f4ce85878f8cca6c4b4f5ac3bb1723e878ef53a6bbf9cccc355fb909d47"} err="failed to get container status \"14862f4ce85878f8cca6c4b4f5ac3bb1723e878ef53a6bbf9cccc355fb909d47\": rpc error: code = NotFound desc = could not find container \"14862f4ce85878f8cca6c4b4f5ac3bb1723e878ef53a6bbf9cccc355fb909d47\": container with ID starting with 14862f4ce85878f8cca6c4b4f5ac3bb1723e878ef53a6bbf9cccc355fb909d47 not found: ID does not exist" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.666767 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/707990a8-7470-4e49-a762-e2d8f5c9382b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "707990a8-7470-4e49-a762-e2d8f5c9382b" (UID: "707990a8-7470-4e49-a762-e2d8f5c9382b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.694718 4787 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/707990a8-7470-4e49-a762-e2d8f5c9382b-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.694758 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/707990a8-7470-4e49-a762-e2d8f5c9382b-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.694768 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmsxp\" (UniqueName: \"kubernetes.io/projected/707990a8-7470-4e49-a762-e2d8f5c9382b-kube-api-access-fmsxp\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.694776 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/707990a8-7470-4e49-a762-e2d8f5c9382b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.694786 4787 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/707990a8-7470-4e49-a762-e2d8f5c9382b-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.694793 4787 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/707990a8-7470-4e49-a762-e2d8f5c9382b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.796400 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c15f26a3-6250-43c9-822a-974a2f3c0a63-config-data\") pod \"c15f26a3-6250-43c9-822a-974a2f3c0a63\" (UID: \"c15f26a3-6250-43c9-822a-974a2f3c0a63\") " Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.796445 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c15f26a3-6250-43c9-822a-974a2f3c0a63-combined-ca-bundle\") pod \"c15f26a3-6250-43c9-822a-974a2f3c0a63\" (UID: \"c15f26a3-6250-43c9-822a-974a2f3c0a63\") " Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.796683 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdfbk\" (UniqueName: \"kubernetes.io/projected/c15f26a3-6250-43c9-822a-974a2f3c0a63-kube-api-access-qdfbk\") pod \"c15f26a3-6250-43c9-822a-974a2f3c0a63\" (UID: \"c15f26a3-6250-43c9-822a-974a2f3c0a63\") " Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.801038 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c15f26a3-6250-43c9-822a-974a2f3c0a63-kube-api-access-qdfbk" (OuterVolumeSpecName: "kube-api-access-qdfbk") pod "c15f26a3-6250-43c9-822a-974a2f3c0a63" (UID: "c15f26a3-6250-43c9-822a-974a2f3c0a63"). InnerVolumeSpecName "kube-api-access-qdfbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.827062 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c15f26a3-6250-43c9-822a-974a2f3c0a63-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c15f26a3-6250-43c9-822a-974a2f3c0a63" (UID: "c15f26a3-6250-43c9-822a-974a2f3c0a63"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.828338 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c15f26a3-6250-43c9-822a-974a2f3c0a63-config-data" (OuterVolumeSpecName: "config-data") pod "c15f26a3-6250-43c9-822a-974a2f3c0a63" (UID: "c15f26a3-6250-43c9-822a-974a2f3c0a63"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.899444 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c15f26a3-6250-43c9-822a-974a2f3c0a63-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.899482 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdfbk\" (UniqueName: \"kubernetes.io/projected/c15f26a3-6250-43c9-822a-974a2f3c0a63-kube-api-access-qdfbk\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.899497 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c15f26a3-6250-43c9-822a-974a2f3c0a63-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.932126 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.940690 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.956728 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 01 09:55:14 crc kubenswrapper[4787]: E1001 09:55:14.957584 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5d060f6-9fac-4c41-b365-033b09e68016" containerName="nova-manage" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.957705 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5d060f6-9fac-4c41-b365-033b09e68016" containerName="nova-manage" Oct 01 09:55:14 crc kubenswrapper[4787]: E1001 09:55:14.957804 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9574c0c-29bf-43bc-b090-35ece3a54e94" containerName="init" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.957891 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9574c0c-29bf-43bc-b090-35ece3a54e94" containerName="init" Oct 01 09:55:14 crc kubenswrapper[4787]: E1001 09:55:14.957983 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="707990a8-7470-4e49-a762-e2d8f5c9382b" containerName="nova-api-log" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.958094 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="707990a8-7470-4e49-a762-e2d8f5c9382b" containerName="nova-api-log" Oct 01 09:55:14 crc kubenswrapper[4787]: E1001 09:55:14.958186 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9574c0c-29bf-43bc-b090-35ece3a54e94" containerName="dnsmasq-dns" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.958280 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9574c0c-29bf-43bc-b090-35ece3a54e94" containerName="dnsmasq-dns" Oct 01 09:55:14 crc kubenswrapper[4787]: E1001 09:55:14.958369 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="707990a8-7470-4e49-a762-e2d8f5c9382b" containerName="nova-api-api" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.958449 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="707990a8-7470-4e49-a762-e2d8f5c9382b" containerName="nova-api-api" Oct 01 09:55:14 crc kubenswrapper[4787]: E1001 09:55:14.958529 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c15f26a3-6250-43c9-822a-974a2f3c0a63" containerName="nova-scheduler-scheduler" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.958616 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c15f26a3-6250-43c9-822a-974a2f3c0a63" containerName="nova-scheduler-scheduler" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.958937 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5d060f6-9fac-4c41-b365-033b09e68016" containerName="nova-manage" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.959043 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="707990a8-7470-4e49-a762-e2d8f5c9382b" containerName="nova-api-log" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.959163 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="707990a8-7470-4e49-a762-e2d8f5c9382b" containerName="nova-api-api" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.959250 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9574c0c-29bf-43bc-b090-35ece3a54e94" containerName="dnsmasq-dns" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.959334 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="c15f26a3-6250-43c9-822a-974a2f3c0a63" containerName="nova-scheduler-scheduler" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.960671 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.965006 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.965020 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.965063 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 01 09:55:14 crc kubenswrapper[4787]: I1001 09:55:14.967730 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 09:55:15 crc kubenswrapper[4787]: I1001 09:55:15.102800 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/59257c48-7bb8-4aad-9f6c-71d4459f2fb7-public-tls-certs\") pod \"nova-api-0\" (UID: \"59257c48-7bb8-4aad-9f6c-71d4459f2fb7\") " pod="openstack/nova-api-0" Oct 01 09:55:15 crc kubenswrapper[4787]: I1001 09:55:15.102896 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59257c48-7bb8-4aad-9f6c-71d4459f2fb7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"59257c48-7bb8-4aad-9f6c-71d4459f2fb7\") " pod="openstack/nova-api-0" Oct 01 09:55:15 crc kubenswrapper[4787]: I1001 09:55:15.102940 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59257c48-7bb8-4aad-9f6c-71d4459f2fb7-logs\") pod \"nova-api-0\" (UID: \"59257c48-7bb8-4aad-9f6c-71d4459f2fb7\") " pod="openstack/nova-api-0" Oct 01 09:55:15 crc kubenswrapper[4787]: I1001 09:55:15.102981 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59257c48-7bb8-4aad-9f6c-71d4459f2fb7-config-data\") pod \"nova-api-0\" (UID: \"59257c48-7bb8-4aad-9f6c-71d4459f2fb7\") " pod="openstack/nova-api-0" Oct 01 09:55:15 crc kubenswrapper[4787]: I1001 09:55:15.103053 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/59257c48-7bb8-4aad-9f6c-71d4459f2fb7-internal-tls-certs\") pod \"nova-api-0\" (UID: \"59257c48-7bb8-4aad-9f6c-71d4459f2fb7\") " pod="openstack/nova-api-0" Oct 01 09:55:15 crc kubenswrapper[4787]: I1001 09:55:15.103110 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrnp7\" (UniqueName: \"kubernetes.io/projected/59257c48-7bb8-4aad-9f6c-71d4459f2fb7-kube-api-access-xrnp7\") pod \"nova-api-0\" (UID: \"59257c48-7bb8-4aad-9f6c-71d4459f2fb7\") " pod="openstack/nova-api-0" Oct 01 09:55:15 crc kubenswrapper[4787]: I1001 09:55:15.205350 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrnp7\" (UniqueName: \"kubernetes.io/projected/59257c48-7bb8-4aad-9f6c-71d4459f2fb7-kube-api-access-xrnp7\") pod \"nova-api-0\" (UID: \"59257c48-7bb8-4aad-9f6c-71d4459f2fb7\") " pod="openstack/nova-api-0" Oct 01 09:55:15 crc kubenswrapper[4787]: I1001 09:55:15.205493 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/59257c48-7bb8-4aad-9f6c-71d4459f2fb7-public-tls-certs\") pod \"nova-api-0\" (UID: \"59257c48-7bb8-4aad-9f6c-71d4459f2fb7\") " pod="openstack/nova-api-0" Oct 01 09:55:15 crc kubenswrapper[4787]: I1001 09:55:15.205568 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59257c48-7bb8-4aad-9f6c-71d4459f2fb7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"59257c48-7bb8-4aad-9f6c-71d4459f2fb7\") " pod="openstack/nova-api-0" Oct 01 09:55:15 crc kubenswrapper[4787]: I1001 09:55:15.205613 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59257c48-7bb8-4aad-9f6c-71d4459f2fb7-logs\") pod \"nova-api-0\" (UID: \"59257c48-7bb8-4aad-9f6c-71d4459f2fb7\") " pod="openstack/nova-api-0" Oct 01 09:55:15 crc kubenswrapper[4787]: I1001 09:55:15.205650 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59257c48-7bb8-4aad-9f6c-71d4459f2fb7-config-data\") pod \"nova-api-0\" (UID: \"59257c48-7bb8-4aad-9f6c-71d4459f2fb7\") " pod="openstack/nova-api-0" Oct 01 09:55:15 crc kubenswrapper[4787]: I1001 09:55:15.205718 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/59257c48-7bb8-4aad-9f6c-71d4459f2fb7-internal-tls-certs\") pod \"nova-api-0\" (UID: \"59257c48-7bb8-4aad-9f6c-71d4459f2fb7\") " pod="openstack/nova-api-0" Oct 01 09:55:15 crc kubenswrapper[4787]: I1001 09:55:15.206174 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59257c48-7bb8-4aad-9f6c-71d4459f2fb7-logs\") pod \"nova-api-0\" (UID: \"59257c48-7bb8-4aad-9f6c-71d4459f2fb7\") " pod="openstack/nova-api-0" Oct 01 09:55:15 crc kubenswrapper[4787]: I1001 09:55:15.209383 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/59257c48-7bb8-4aad-9f6c-71d4459f2fb7-internal-tls-certs\") pod \"nova-api-0\" (UID: \"59257c48-7bb8-4aad-9f6c-71d4459f2fb7\") " pod="openstack/nova-api-0" Oct 01 09:55:15 crc kubenswrapper[4787]: I1001 09:55:15.210174 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/59257c48-7bb8-4aad-9f6c-71d4459f2fb7-public-tls-certs\") pod \"nova-api-0\" (UID: \"59257c48-7bb8-4aad-9f6c-71d4459f2fb7\") " pod="openstack/nova-api-0" Oct 01 09:55:15 crc kubenswrapper[4787]: I1001 09:55:15.212685 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59257c48-7bb8-4aad-9f6c-71d4459f2fb7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"59257c48-7bb8-4aad-9f6c-71d4459f2fb7\") " pod="openstack/nova-api-0" Oct 01 09:55:15 crc kubenswrapper[4787]: I1001 09:55:15.212730 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59257c48-7bb8-4aad-9f6c-71d4459f2fb7-config-data\") pod \"nova-api-0\" (UID: \"59257c48-7bb8-4aad-9f6c-71d4459f2fb7\") " pod="openstack/nova-api-0" Oct 01 09:55:15 crc kubenswrapper[4787]: I1001 09:55:15.222647 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrnp7\" (UniqueName: \"kubernetes.io/projected/59257c48-7bb8-4aad-9f6c-71d4459f2fb7-kube-api-access-xrnp7\") pod \"nova-api-0\" (UID: \"59257c48-7bb8-4aad-9f6c-71d4459f2fb7\") " pod="openstack/nova-api-0" Oct 01 09:55:15 crc kubenswrapper[4787]: I1001 09:55:15.290325 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 01 09:55:15 crc kubenswrapper[4787]: I1001 09:55:15.601348 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 09:55:15 crc kubenswrapper[4787]: I1001 09:55:15.601382 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c15f26a3-6250-43c9-822a-974a2f3c0a63","Type":"ContainerDied","Data":"ea86e3d2a7445df104cd9e94669cebbdf1db437a974770aabfcde27ec96380bc"} Oct 01 09:55:15 crc kubenswrapper[4787]: I1001 09:55:15.601815 4787 scope.go:117] "RemoveContainer" containerID="e1033e70a9ecb6214b67b36775d24b8575db35d5df073e0b8ff0126db0442282" Oct 01 09:55:15 crc kubenswrapper[4787]: I1001 09:55:15.644087 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 09:55:15 crc kubenswrapper[4787]: I1001 09:55:15.660991 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 09:55:15 crc kubenswrapper[4787]: I1001 09:55:15.685917 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 09:55:15 crc kubenswrapper[4787]: I1001 09:55:15.687315 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 09:55:15 crc kubenswrapper[4787]: I1001 09:55:15.692715 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 01 09:55:15 crc kubenswrapper[4787]: I1001 09:55:15.695796 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 09:55:15 crc kubenswrapper[4787]: I1001 09:55:15.774641 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 01 09:55:15 crc kubenswrapper[4787]: W1001 09:55:15.782547 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod59257c48_7bb8_4aad_9f6c_71d4459f2fb7.slice/crio-baab7a15e60d8742f9798651de75d09e1d2f7d01ffbc95bc74c0f3ef58d11a16 WatchSource:0}: Error finding container baab7a15e60d8742f9798651de75d09e1d2f7d01ffbc95bc74c0f3ef58d11a16: Status 404 returned error can't find the container with id baab7a15e60d8742f9798651de75d09e1d2f7d01ffbc95bc74c0f3ef58d11a16 Oct 01 09:55:15 crc kubenswrapper[4787]: I1001 09:55:15.816379 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35a5073c-d1ab-4fa9-b1c5-3a6df1d42bcc-config-data\") pod \"nova-scheduler-0\" (UID: \"35a5073c-d1ab-4fa9-b1c5-3a6df1d42bcc\") " pod="openstack/nova-scheduler-0" Oct 01 09:55:15 crc kubenswrapper[4787]: I1001 09:55:15.816483 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35a5073c-d1ab-4fa9-b1c5-3a6df1d42bcc-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"35a5073c-d1ab-4fa9-b1c5-3a6df1d42bcc\") " pod="openstack/nova-scheduler-0" Oct 01 09:55:15 crc kubenswrapper[4787]: I1001 09:55:15.816514 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kkhv\" (UniqueName: \"kubernetes.io/projected/35a5073c-d1ab-4fa9-b1c5-3a6df1d42bcc-kube-api-access-6kkhv\") pod \"nova-scheduler-0\" (UID: \"35a5073c-d1ab-4fa9-b1c5-3a6df1d42bcc\") " pod="openstack/nova-scheduler-0" Oct 01 09:55:15 crc kubenswrapper[4787]: I1001 09:55:15.918618 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35a5073c-d1ab-4fa9-b1c5-3a6df1d42bcc-config-data\") pod \"nova-scheduler-0\" (UID: \"35a5073c-d1ab-4fa9-b1c5-3a6df1d42bcc\") " pod="openstack/nova-scheduler-0" Oct 01 09:55:15 crc kubenswrapper[4787]: I1001 09:55:15.918701 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35a5073c-d1ab-4fa9-b1c5-3a6df1d42bcc-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"35a5073c-d1ab-4fa9-b1c5-3a6df1d42bcc\") " pod="openstack/nova-scheduler-0" Oct 01 09:55:15 crc kubenswrapper[4787]: I1001 09:55:15.918731 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kkhv\" (UniqueName: \"kubernetes.io/projected/35a5073c-d1ab-4fa9-b1c5-3a6df1d42bcc-kube-api-access-6kkhv\") pod \"nova-scheduler-0\" (UID: \"35a5073c-d1ab-4fa9-b1c5-3a6df1d42bcc\") " pod="openstack/nova-scheduler-0" Oct 01 09:55:15 crc kubenswrapper[4787]: I1001 09:55:15.924449 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35a5073c-d1ab-4fa9-b1c5-3a6df1d42bcc-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"35a5073c-d1ab-4fa9-b1c5-3a6df1d42bcc\") " pod="openstack/nova-scheduler-0" Oct 01 09:55:15 crc kubenswrapper[4787]: I1001 09:55:15.925044 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35a5073c-d1ab-4fa9-b1c5-3a6df1d42bcc-config-data\") pod \"nova-scheduler-0\" (UID: \"35a5073c-d1ab-4fa9-b1c5-3a6df1d42bcc\") " pod="openstack/nova-scheduler-0" Oct 01 09:55:15 crc kubenswrapper[4787]: I1001 09:55:15.940092 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kkhv\" (UniqueName: \"kubernetes.io/projected/35a5073c-d1ab-4fa9-b1c5-3a6df1d42bcc-kube-api-access-6kkhv\") pod \"nova-scheduler-0\" (UID: \"35a5073c-d1ab-4fa9-b1c5-3a6df1d42bcc\") " pod="openstack/nova-scheduler-0" Oct 01 09:55:16 crc kubenswrapper[4787]: I1001 09:55:16.005641 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 01 09:55:16 crc kubenswrapper[4787]: I1001 09:55:16.547094 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="707990a8-7470-4e49-a762-e2d8f5c9382b" path="/var/lib/kubelet/pods/707990a8-7470-4e49-a762-e2d8f5c9382b/volumes" Oct 01 09:55:16 crc kubenswrapper[4787]: I1001 09:55:16.548465 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c15f26a3-6250-43c9-822a-974a2f3c0a63" path="/var/lib/kubelet/pods/c15f26a3-6250-43c9-822a-974a2f3c0a63/volumes" Oct 01 09:55:16 crc kubenswrapper[4787]: I1001 09:55:16.551161 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 01 09:55:16 crc kubenswrapper[4787]: W1001 09:55:16.560041 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35a5073c_d1ab_4fa9_b1c5_3a6df1d42bcc.slice/crio-d9735cbeefb87e7b86f7d96bc0735b8321285042757d47e5889e23a1110dfde1 WatchSource:0}: Error finding container d9735cbeefb87e7b86f7d96bc0735b8321285042757d47e5889e23a1110dfde1: Status 404 returned error can't find the container with id d9735cbeefb87e7b86f7d96bc0735b8321285042757d47e5889e23a1110dfde1 Oct 01 09:55:16 crc kubenswrapper[4787]: I1001 09:55:16.614776 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"35a5073c-d1ab-4fa9-b1c5-3a6df1d42bcc","Type":"ContainerStarted","Data":"d9735cbeefb87e7b86f7d96bc0735b8321285042757d47e5889e23a1110dfde1"} Oct 01 09:55:16 crc kubenswrapper[4787]: I1001 09:55:16.617279 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"59257c48-7bb8-4aad-9f6c-71d4459f2fb7","Type":"ContainerStarted","Data":"c796a42ec2604657ceb0d725b3d844067f5c40b96d0fd4d9e2db72c0ec267911"} Oct 01 09:55:16 crc kubenswrapper[4787]: I1001 09:55:16.617328 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"59257c48-7bb8-4aad-9f6c-71d4459f2fb7","Type":"ContainerStarted","Data":"d2fc9654191c47099a3a34daf3224748190025054b25f88f0d02bbc83c3e8a65"} Oct 01 09:55:16 crc kubenswrapper[4787]: I1001 09:55:16.617339 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"59257c48-7bb8-4aad-9f6c-71d4459f2fb7","Type":"ContainerStarted","Data":"baab7a15e60d8742f9798651de75d09e1d2f7d01ffbc95bc74c0f3ef58d11a16"} Oct 01 09:55:16 crc kubenswrapper[4787]: I1001 09:55:16.657419 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.657400433 podStartE2EDuration="2.657400433s" podCreationTimestamp="2025-10-01 09:55:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:55:16.635460356 +0000 UTC m=+1148.750604523" watchObservedRunningTime="2025-10-01 09:55:16.657400433 +0000 UTC m=+1148.772544590" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.355986 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.466015 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/700e3732-7195-48fb-a5f6-4fb80b9f13e8-logs\") pod \"700e3732-7195-48fb-a5f6-4fb80b9f13e8\" (UID: \"700e3732-7195-48fb-a5f6-4fb80b9f13e8\") " Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.466138 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/700e3732-7195-48fb-a5f6-4fb80b9f13e8-combined-ca-bundle\") pod \"700e3732-7195-48fb-a5f6-4fb80b9f13e8\" (UID: \"700e3732-7195-48fb-a5f6-4fb80b9f13e8\") " Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.466350 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/700e3732-7195-48fb-a5f6-4fb80b9f13e8-config-data\") pod \"700e3732-7195-48fb-a5f6-4fb80b9f13e8\" (UID: \"700e3732-7195-48fb-a5f6-4fb80b9f13e8\") " Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.466383 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hzd4v\" (UniqueName: \"kubernetes.io/projected/700e3732-7195-48fb-a5f6-4fb80b9f13e8-kube-api-access-hzd4v\") pod \"700e3732-7195-48fb-a5f6-4fb80b9f13e8\" (UID: \"700e3732-7195-48fb-a5f6-4fb80b9f13e8\") " Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.466421 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/700e3732-7195-48fb-a5f6-4fb80b9f13e8-nova-metadata-tls-certs\") pod \"700e3732-7195-48fb-a5f6-4fb80b9f13e8\" (UID: \"700e3732-7195-48fb-a5f6-4fb80b9f13e8\") " Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.466893 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/700e3732-7195-48fb-a5f6-4fb80b9f13e8-logs" (OuterVolumeSpecName: "logs") pod "700e3732-7195-48fb-a5f6-4fb80b9f13e8" (UID: "700e3732-7195-48fb-a5f6-4fb80b9f13e8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.467166 4787 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/700e3732-7195-48fb-a5f6-4fb80b9f13e8-logs\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.474972 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/700e3732-7195-48fb-a5f6-4fb80b9f13e8-kube-api-access-hzd4v" (OuterVolumeSpecName: "kube-api-access-hzd4v") pod "700e3732-7195-48fb-a5f6-4fb80b9f13e8" (UID: "700e3732-7195-48fb-a5f6-4fb80b9f13e8"). InnerVolumeSpecName "kube-api-access-hzd4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.494748 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/700e3732-7195-48fb-a5f6-4fb80b9f13e8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "700e3732-7195-48fb-a5f6-4fb80b9f13e8" (UID: "700e3732-7195-48fb-a5f6-4fb80b9f13e8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.497206 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/700e3732-7195-48fb-a5f6-4fb80b9f13e8-config-data" (OuterVolumeSpecName: "config-data") pod "700e3732-7195-48fb-a5f6-4fb80b9f13e8" (UID: "700e3732-7195-48fb-a5f6-4fb80b9f13e8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.536818 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/700e3732-7195-48fb-a5f6-4fb80b9f13e8-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "700e3732-7195-48fb-a5f6-4fb80b9f13e8" (UID: "700e3732-7195-48fb-a5f6-4fb80b9f13e8"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.569227 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/700e3732-7195-48fb-a5f6-4fb80b9f13e8-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.569260 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hzd4v\" (UniqueName: \"kubernetes.io/projected/700e3732-7195-48fb-a5f6-4fb80b9f13e8-kube-api-access-hzd4v\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.569275 4787 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/700e3732-7195-48fb-a5f6-4fb80b9f13e8-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.569287 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/700e3732-7195-48fb-a5f6-4fb80b9f13e8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.631444 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"35a5073c-d1ab-4fa9-b1c5-3a6df1d42bcc","Type":"ContainerStarted","Data":"41578764bd2ee1cca10c1361d0a038615f4a4f194faafcb9f561248ef3f7aff0"} Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.636102 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"700e3732-7195-48fb-a5f6-4fb80b9f13e8","Type":"ContainerDied","Data":"9a4e6b2ebf2a9aee84a3e28a27294b0ceba00c22a18adce42fe141eb75395696"} Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.636146 4787 scope.go:117] "RemoveContainer" containerID="9a4e6b2ebf2a9aee84a3e28a27294b0ceba00c22a18adce42fe141eb75395696" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.636123 4787 generic.go:334] "Generic (PLEG): container finished" podID="700e3732-7195-48fb-a5f6-4fb80b9f13e8" containerID="9a4e6b2ebf2a9aee84a3e28a27294b0ceba00c22a18adce42fe141eb75395696" exitCode=0 Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.636260 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"700e3732-7195-48fb-a5f6-4fb80b9f13e8","Type":"ContainerDied","Data":"08d31b9e424648c555b249db62424526da244f58e3101e65ff6d91d738ad8729"} Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.636557 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.657346 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.65730955 podStartE2EDuration="2.65730955s" podCreationTimestamp="2025-10-01 09:55:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:55:17.653181606 +0000 UTC m=+1149.768325773" watchObservedRunningTime="2025-10-01 09:55:17.65730955 +0000 UTC m=+1149.772453757" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.661647 4787 scope.go:117] "RemoveContainer" containerID="5a17511d23a3c2c840e42020b72de1cd5076c84cd25e857681d078bc8332adca" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.683962 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.692374 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.717835 4787 scope.go:117] "RemoveContainer" containerID="9a4e6b2ebf2a9aee84a3e28a27294b0ceba00c22a18adce42fe141eb75395696" Oct 01 09:55:17 crc kubenswrapper[4787]: E1001 09:55:17.718483 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a4e6b2ebf2a9aee84a3e28a27294b0ceba00c22a18adce42fe141eb75395696\": container with ID starting with 9a4e6b2ebf2a9aee84a3e28a27294b0ceba00c22a18adce42fe141eb75395696 not found: ID does not exist" containerID="9a4e6b2ebf2a9aee84a3e28a27294b0ceba00c22a18adce42fe141eb75395696" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.718514 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a4e6b2ebf2a9aee84a3e28a27294b0ceba00c22a18adce42fe141eb75395696"} err="failed to get container status \"9a4e6b2ebf2a9aee84a3e28a27294b0ceba00c22a18adce42fe141eb75395696\": rpc error: code = NotFound desc = could not find container \"9a4e6b2ebf2a9aee84a3e28a27294b0ceba00c22a18adce42fe141eb75395696\": container with ID starting with 9a4e6b2ebf2a9aee84a3e28a27294b0ceba00c22a18adce42fe141eb75395696 not found: ID does not exist" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.718536 4787 scope.go:117] "RemoveContainer" containerID="5a17511d23a3c2c840e42020b72de1cd5076c84cd25e857681d078bc8332adca" Oct 01 09:55:17 crc kubenswrapper[4787]: E1001 09:55:17.718805 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a17511d23a3c2c840e42020b72de1cd5076c84cd25e857681d078bc8332adca\": container with ID starting with 5a17511d23a3c2c840e42020b72de1cd5076c84cd25e857681d078bc8332adca not found: ID does not exist" containerID="5a17511d23a3c2c840e42020b72de1cd5076c84cd25e857681d078bc8332adca" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.718827 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a17511d23a3c2c840e42020b72de1cd5076c84cd25e857681d078bc8332adca"} err="failed to get container status \"5a17511d23a3c2c840e42020b72de1cd5076c84cd25e857681d078bc8332adca\": rpc error: code = NotFound desc = could not find container \"5a17511d23a3c2c840e42020b72de1cd5076c84cd25e857681d078bc8332adca\": container with ID starting with 5a17511d23a3c2c840e42020b72de1cd5076c84cd25e857681d078bc8332adca not found: ID does not exist" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.732308 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:55:17 crc kubenswrapper[4787]: E1001 09:55:17.733019 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="700e3732-7195-48fb-a5f6-4fb80b9f13e8" containerName="nova-metadata-metadata" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.733042 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="700e3732-7195-48fb-a5f6-4fb80b9f13e8" containerName="nova-metadata-metadata" Oct 01 09:55:17 crc kubenswrapper[4787]: E1001 09:55:17.733069 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="700e3732-7195-48fb-a5f6-4fb80b9f13e8" containerName="nova-metadata-log" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.733098 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="700e3732-7195-48fb-a5f6-4fb80b9f13e8" containerName="nova-metadata-log" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.733307 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="700e3732-7195-48fb-a5f6-4fb80b9f13e8" containerName="nova-metadata-log" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.733325 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="700e3732-7195-48fb-a5f6-4fb80b9f13e8" containerName="nova-metadata-metadata" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.734478 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.742574 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.742808 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.743925 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.876876 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3dac849-090f-4eaa-bd8a-11a9e85f0317-config-data\") pod \"nova-metadata-0\" (UID: \"f3dac849-090f-4eaa-bd8a-11a9e85f0317\") " pod="openstack/nova-metadata-0" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.877029 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rl5nr\" (UniqueName: \"kubernetes.io/projected/f3dac849-090f-4eaa-bd8a-11a9e85f0317-kube-api-access-rl5nr\") pod \"nova-metadata-0\" (UID: \"f3dac849-090f-4eaa-bd8a-11a9e85f0317\") " pod="openstack/nova-metadata-0" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.877161 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3dac849-090f-4eaa-bd8a-11a9e85f0317-logs\") pod \"nova-metadata-0\" (UID: \"f3dac849-090f-4eaa-bd8a-11a9e85f0317\") " pod="openstack/nova-metadata-0" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.877194 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3dac849-090f-4eaa-bd8a-11a9e85f0317-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f3dac849-090f-4eaa-bd8a-11a9e85f0317\") " pod="openstack/nova-metadata-0" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.877254 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3dac849-090f-4eaa-bd8a-11a9e85f0317-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f3dac849-090f-4eaa-bd8a-11a9e85f0317\") " pod="openstack/nova-metadata-0" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.979898 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3dac849-090f-4eaa-bd8a-11a9e85f0317-config-data\") pod \"nova-metadata-0\" (UID: \"f3dac849-090f-4eaa-bd8a-11a9e85f0317\") " pod="openstack/nova-metadata-0" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.980033 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rl5nr\" (UniqueName: \"kubernetes.io/projected/f3dac849-090f-4eaa-bd8a-11a9e85f0317-kube-api-access-rl5nr\") pod \"nova-metadata-0\" (UID: \"f3dac849-090f-4eaa-bd8a-11a9e85f0317\") " pod="openstack/nova-metadata-0" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.980172 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3dac849-090f-4eaa-bd8a-11a9e85f0317-logs\") pod \"nova-metadata-0\" (UID: \"f3dac849-090f-4eaa-bd8a-11a9e85f0317\") " pod="openstack/nova-metadata-0" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.980240 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3dac849-090f-4eaa-bd8a-11a9e85f0317-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f3dac849-090f-4eaa-bd8a-11a9e85f0317\") " pod="openstack/nova-metadata-0" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.980352 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3dac849-090f-4eaa-bd8a-11a9e85f0317-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f3dac849-090f-4eaa-bd8a-11a9e85f0317\") " pod="openstack/nova-metadata-0" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.981105 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3dac849-090f-4eaa-bd8a-11a9e85f0317-logs\") pod \"nova-metadata-0\" (UID: \"f3dac849-090f-4eaa-bd8a-11a9e85f0317\") " pod="openstack/nova-metadata-0" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.990212 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3dac849-090f-4eaa-bd8a-11a9e85f0317-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f3dac849-090f-4eaa-bd8a-11a9e85f0317\") " pod="openstack/nova-metadata-0" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.990473 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3dac849-090f-4eaa-bd8a-11a9e85f0317-config-data\") pod \"nova-metadata-0\" (UID: \"f3dac849-090f-4eaa-bd8a-11a9e85f0317\") " pod="openstack/nova-metadata-0" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.992868 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3dac849-090f-4eaa-bd8a-11a9e85f0317-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f3dac849-090f-4eaa-bd8a-11a9e85f0317\") " pod="openstack/nova-metadata-0" Oct 01 09:55:17 crc kubenswrapper[4787]: I1001 09:55:17.999209 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rl5nr\" (UniqueName: \"kubernetes.io/projected/f3dac849-090f-4eaa-bd8a-11a9e85f0317-kube-api-access-rl5nr\") pod \"nova-metadata-0\" (UID: \"f3dac849-090f-4eaa-bd8a-11a9e85f0317\") " pod="openstack/nova-metadata-0" Oct 01 09:55:18 crc kubenswrapper[4787]: I1001 09:55:18.061395 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 01 09:55:18 crc kubenswrapper[4787]: I1001 09:55:18.537297 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="700e3732-7195-48fb-a5f6-4fb80b9f13e8" path="/var/lib/kubelet/pods/700e3732-7195-48fb-a5f6-4fb80b9f13e8/volumes" Oct 01 09:55:18 crc kubenswrapper[4787]: I1001 09:55:18.545202 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 01 09:55:18 crc kubenswrapper[4787]: I1001 09:55:18.653560 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f3dac849-090f-4eaa-bd8a-11a9e85f0317","Type":"ContainerStarted","Data":"aa549b9994a175addf27b9aee6f2c1afe2b678730f410844dcc3b425c4f5d007"} Oct 01 09:55:19 crc kubenswrapper[4787]: I1001 09:55:19.670363 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f3dac849-090f-4eaa-bd8a-11a9e85f0317","Type":"ContainerStarted","Data":"3cedd6f235d246274e98645302330cb150ff0bffd0f6033dab8369dd256f7b3a"} Oct 01 09:55:19 crc kubenswrapper[4787]: I1001 09:55:19.670802 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f3dac849-090f-4eaa-bd8a-11a9e85f0317","Type":"ContainerStarted","Data":"c63efc57ddf9a1b6434f7efe35fe4c0aae60361f58de017cfe4646969270d33e"} Oct 01 09:55:19 crc kubenswrapper[4787]: I1001 09:55:19.699585 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.699555762 podStartE2EDuration="2.699555762s" podCreationTimestamp="2025-10-01 09:55:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:55:19.685513486 +0000 UTC m=+1151.800657663" watchObservedRunningTime="2025-10-01 09:55:19.699555762 +0000 UTC m=+1151.814699939" Oct 01 09:55:21 crc kubenswrapper[4787]: I1001 09:55:21.006656 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 01 09:55:23 crc kubenswrapper[4787]: I1001 09:55:23.063105 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 01 09:55:23 crc kubenswrapper[4787]: I1001 09:55:23.066456 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 01 09:55:25 crc kubenswrapper[4787]: I1001 09:55:25.291700 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 01 09:55:25 crc kubenswrapper[4787]: I1001 09:55:25.292591 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 01 09:55:26 crc kubenswrapper[4787]: I1001 09:55:26.006640 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 01 09:55:26 crc kubenswrapper[4787]: I1001 09:55:26.068304 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 01 09:55:26 crc kubenswrapper[4787]: I1001 09:55:26.304341 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="59257c48-7bb8-4aad-9f6c-71d4459f2fb7" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.205:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 09:55:26 crc kubenswrapper[4787]: I1001 09:55:26.304385 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="59257c48-7bb8-4aad-9f6c-71d4459f2fb7" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.205:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 09:55:26 crc kubenswrapper[4787]: I1001 09:55:26.800990 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 01 09:55:28 crc kubenswrapper[4787]: I1001 09:55:28.061762 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 01 09:55:28 crc kubenswrapper[4787]: I1001 09:55:28.063452 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 01 09:55:29 crc kubenswrapper[4787]: I1001 09:55:29.076464 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f3dac849-090f-4eaa-bd8a-11a9e85f0317" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.207:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 09:55:29 crc kubenswrapper[4787]: I1001 09:55:29.076467 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f3dac849-090f-4eaa-bd8a-11a9e85f0317" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.207:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 01 09:55:32 crc kubenswrapper[4787]: I1001 09:55:32.825924 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 01 09:55:35 crc kubenswrapper[4787]: I1001 09:55:35.297825 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 01 09:55:35 crc kubenswrapper[4787]: I1001 09:55:35.299298 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 01 09:55:35 crc kubenswrapper[4787]: I1001 09:55:35.299346 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 01 09:55:35 crc kubenswrapper[4787]: I1001 09:55:35.306844 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 01 09:55:35 crc kubenswrapper[4787]: I1001 09:55:35.866694 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 01 09:55:35 crc kubenswrapper[4787]: I1001 09:55:35.875538 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 01 09:55:38 crc kubenswrapper[4787]: I1001 09:55:38.066705 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 01 09:55:38 crc kubenswrapper[4787]: I1001 09:55:38.068887 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 01 09:55:38 crc kubenswrapper[4787]: I1001 09:55:38.071044 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 01 09:55:38 crc kubenswrapper[4787]: I1001 09:55:38.902631 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 01 09:55:47 crc kubenswrapper[4787]: I1001 09:55:47.000003 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 09:55:47 crc kubenswrapper[4787]: I1001 09:55:47.908809 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 09:55:51 crc kubenswrapper[4787]: I1001 09:55:51.586772 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="e762572f-d5b9-462e-9953-7143f648c9ae" containerName="rabbitmq" containerID="cri-o://dc3a7667b021a77fbd7004c14e7eea8bcaffddf0b1788a0efc6e5c22d34ef255" gracePeriod=604796 Oct 01 09:55:52 crc kubenswrapper[4787]: I1001 09:55:52.334042 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="ce587847-47c2-41de-95d1-c9f8ab88961e" containerName="rabbitmq" containerID="cri-o://7fc382e561b660b35dc13b61c6cea44e5887caaef5da418260ae272e6cffb258" gracePeriod=604796 Oct 01 09:55:57 crc kubenswrapper[4787]: I1001 09:55:57.889191 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="e762572f-d5b9-462e-9953-7143f648c9ae" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.100:5671: connect: connection refused" Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.161849 4787 generic.go:334] "Generic (PLEG): container finished" podID="e762572f-d5b9-462e-9953-7143f648c9ae" containerID="dc3a7667b021a77fbd7004c14e7eea8bcaffddf0b1788a0efc6e5c22d34ef255" exitCode=0 Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.161919 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e762572f-d5b9-462e-9953-7143f648c9ae","Type":"ContainerDied","Data":"dc3a7667b021a77fbd7004c14e7eea8bcaffddf0b1788a0efc6e5c22d34ef255"} Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.161970 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e762572f-d5b9-462e-9953-7143f648c9ae","Type":"ContainerDied","Data":"1fa050ebea47947b00534e1bcd8a7e26894e48cad6c5ba6f3fba7bfb5bc0d198"} Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.161988 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1fa050ebea47947b00534e1bcd8a7e26894e48cad6c5ba6f3fba7bfb5bc0d198" Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.178029 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.191412 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e762572f-d5b9-462e-9953-7143f648c9ae-rabbitmq-erlang-cookie\") pod \"e762572f-d5b9-462e-9953-7143f648c9ae\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.191478 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e762572f-d5b9-462e-9953-7143f648c9ae-rabbitmq-confd\") pod \"e762572f-d5b9-462e-9953-7143f648c9ae\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.191541 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e762572f-d5b9-462e-9953-7143f648c9ae-rabbitmq-plugins\") pod \"e762572f-d5b9-462e-9953-7143f648c9ae\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.191588 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e762572f-d5b9-462e-9953-7143f648c9ae-pod-info\") pod \"e762572f-d5b9-462e-9953-7143f648c9ae\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.191620 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e762572f-d5b9-462e-9953-7143f648c9ae-rabbitmq-tls\") pod \"e762572f-d5b9-462e-9953-7143f648c9ae\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.191645 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e762572f-d5b9-462e-9953-7143f648c9ae-config-data\") pod \"e762572f-d5b9-462e-9953-7143f648c9ae\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.191669 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"e762572f-d5b9-462e-9953-7143f648c9ae\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.191714 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e762572f-d5b9-462e-9953-7143f648c9ae-server-conf\") pod \"e762572f-d5b9-462e-9953-7143f648c9ae\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.191771 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e762572f-d5b9-462e-9953-7143f648c9ae-plugins-conf\") pod \"e762572f-d5b9-462e-9953-7143f648c9ae\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.191801 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e762572f-d5b9-462e-9953-7143f648c9ae-erlang-cookie-secret\") pod \"e762572f-d5b9-462e-9953-7143f648c9ae\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.191833 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6qkc\" (UniqueName: \"kubernetes.io/projected/e762572f-d5b9-462e-9953-7143f648c9ae-kube-api-access-c6qkc\") pod \"e762572f-d5b9-462e-9953-7143f648c9ae\" (UID: \"e762572f-d5b9-462e-9953-7143f648c9ae\") " Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.192294 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e762572f-d5b9-462e-9953-7143f648c9ae-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "e762572f-d5b9-462e-9953-7143f648c9ae" (UID: "e762572f-d5b9-462e-9953-7143f648c9ae"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.192722 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e762572f-d5b9-462e-9953-7143f648c9ae-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "e762572f-d5b9-462e-9953-7143f648c9ae" (UID: "e762572f-d5b9-462e-9953-7143f648c9ae"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.194051 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e762572f-d5b9-462e-9953-7143f648c9ae-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "e762572f-d5b9-462e-9953-7143f648c9ae" (UID: "e762572f-d5b9-462e-9953-7143f648c9ae"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.201135 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e762572f-d5b9-462e-9953-7143f648c9ae-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "e762572f-d5b9-462e-9953-7143f648c9ae" (UID: "e762572f-d5b9-462e-9953-7143f648c9ae"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.201202 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e762572f-d5b9-462e-9953-7143f648c9ae-kube-api-access-c6qkc" (OuterVolumeSpecName: "kube-api-access-c6qkc") pod "e762572f-d5b9-462e-9953-7143f648c9ae" (UID: "e762572f-d5b9-462e-9953-7143f648c9ae"). InnerVolumeSpecName "kube-api-access-c6qkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.201803 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/e762572f-d5b9-462e-9953-7143f648c9ae-pod-info" (OuterVolumeSpecName: "pod-info") pod "e762572f-d5b9-462e-9953-7143f648c9ae" (UID: "e762572f-d5b9-462e-9953-7143f648c9ae"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.201818 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "e762572f-d5b9-462e-9953-7143f648c9ae" (UID: "e762572f-d5b9-462e-9953-7143f648c9ae"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.225279 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e762572f-d5b9-462e-9953-7143f648c9ae-config-data" (OuterVolumeSpecName: "config-data") pod "e762572f-d5b9-462e-9953-7143f648c9ae" (UID: "e762572f-d5b9-462e-9953-7143f648c9ae"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.243458 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e762572f-d5b9-462e-9953-7143f648c9ae-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "e762572f-d5b9-462e-9953-7143f648c9ae" (UID: "e762572f-d5b9-462e-9953-7143f648c9ae"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.253632 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="ce587847-47c2-41de-95d1-c9f8ab88961e" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.101:5671: connect: connection refused" Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.294875 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e762572f-d5b9-462e-9953-7143f648c9ae-server-conf" (OuterVolumeSpecName: "server-conf") pod "e762572f-d5b9-462e-9953-7143f648c9ae" (UID: "e762572f-d5b9-462e-9953-7143f648c9ae"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.297153 4787 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e762572f-d5b9-462e-9953-7143f648c9ae-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.298622 4787 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e762572f-d5b9-462e-9953-7143f648c9ae-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.298766 4787 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e762572f-d5b9-462e-9953-7143f648c9ae-pod-info\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.298913 4787 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e762572f-d5b9-462e-9953-7143f648c9ae-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.299022 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e762572f-d5b9-462e-9953-7143f648c9ae-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.299149 4787 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.299856 4787 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e762572f-d5b9-462e-9953-7143f648c9ae-server-conf\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.299975 4787 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e762572f-d5b9-462e-9953-7143f648c9ae-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.300134 4787 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e762572f-d5b9-462e-9953-7143f648c9ae-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.300299 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6qkc\" (UniqueName: \"kubernetes.io/projected/e762572f-d5b9-462e-9953-7143f648c9ae-kube-api-access-c6qkc\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.329376 4787 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.373599 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e762572f-d5b9-462e-9953-7143f648c9ae-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "e762572f-d5b9-462e-9953-7143f648c9ae" (UID: "e762572f-d5b9-462e-9953-7143f648c9ae"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.401802 4787 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.401838 4787 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e762572f-d5b9-462e-9953-7143f648c9ae-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:58 crc kubenswrapper[4787]: E1001 09:55:58.822818 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode762572f_d5b9_462e_9953_7143f648c9ae.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode762572f_d5b9_462e_9953_7143f648c9ae.slice/crio-1fa050ebea47947b00534e1bcd8a7e26894e48cad6c5ba6f3fba7bfb5bc0d198\": RecentStats: unable to find data in memory cache]" Oct 01 09:55:58 crc kubenswrapper[4787]: I1001 09:55:58.965695 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.010962 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ce587847-47c2-41de-95d1-c9f8ab88961e-config-data\") pod \"ce587847-47c2-41de-95d1-c9f8ab88961e\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.011012 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ce587847-47c2-41de-95d1-c9f8ab88961e-erlang-cookie-secret\") pod \"ce587847-47c2-41de-95d1-c9f8ab88961e\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.011043 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqw5k\" (UniqueName: \"kubernetes.io/projected/ce587847-47c2-41de-95d1-c9f8ab88961e-kube-api-access-tqw5k\") pod \"ce587847-47c2-41de-95d1-c9f8ab88961e\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.011099 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ce587847-47c2-41de-95d1-c9f8ab88961e-rabbitmq-confd\") pod \"ce587847-47c2-41de-95d1-c9f8ab88961e\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.011181 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ce587847-47c2-41de-95d1-c9f8ab88961e-rabbitmq-tls\") pod \"ce587847-47c2-41de-95d1-c9f8ab88961e\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.011206 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ce587847-47c2-41de-95d1-c9f8ab88961e\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.011287 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ce587847-47c2-41de-95d1-c9f8ab88961e-rabbitmq-erlang-cookie\") pod \"ce587847-47c2-41de-95d1-c9f8ab88961e\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.011333 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ce587847-47c2-41de-95d1-c9f8ab88961e-rabbitmq-plugins\") pod \"ce587847-47c2-41de-95d1-c9f8ab88961e\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.011383 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ce587847-47c2-41de-95d1-c9f8ab88961e-pod-info\") pod \"ce587847-47c2-41de-95d1-c9f8ab88961e\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.011416 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ce587847-47c2-41de-95d1-c9f8ab88961e-server-conf\") pod \"ce587847-47c2-41de-95d1-c9f8ab88961e\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.011455 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ce587847-47c2-41de-95d1-c9f8ab88961e-plugins-conf\") pod \"ce587847-47c2-41de-95d1-c9f8ab88961e\" (UID: \"ce587847-47c2-41de-95d1-c9f8ab88961e\") " Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.012447 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce587847-47c2-41de-95d1-c9f8ab88961e-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "ce587847-47c2-41de-95d1-c9f8ab88961e" (UID: "ce587847-47c2-41de-95d1-c9f8ab88961e"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.018197 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce587847-47c2-41de-95d1-c9f8ab88961e-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "ce587847-47c2-41de-95d1-c9f8ab88961e" (UID: "ce587847-47c2-41de-95d1-c9f8ab88961e"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.018516 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce587847-47c2-41de-95d1-c9f8ab88961e-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "ce587847-47c2-41de-95d1-c9f8ab88961e" (UID: "ce587847-47c2-41de-95d1-c9f8ab88961e"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.021814 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce587847-47c2-41de-95d1-c9f8ab88961e-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "ce587847-47c2-41de-95d1-c9f8ab88961e" (UID: "ce587847-47c2-41de-95d1-c9f8ab88961e"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.021866 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce587847-47c2-41de-95d1-c9f8ab88961e-kube-api-access-tqw5k" (OuterVolumeSpecName: "kube-api-access-tqw5k") pod "ce587847-47c2-41de-95d1-c9f8ab88961e" (UID: "ce587847-47c2-41de-95d1-c9f8ab88961e"). InnerVolumeSpecName "kube-api-access-tqw5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.025248 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce587847-47c2-41de-95d1-c9f8ab88961e-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "ce587847-47c2-41de-95d1-c9f8ab88961e" (UID: "ce587847-47c2-41de-95d1-c9f8ab88961e"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.027134 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "ce587847-47c2-41de-95d1-c9f8ab88961e" (UID: "ce587847-47c2-41de-95d1-c9f8ab88961e"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.030470 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/ce587847-47c2-41de-95d1-c9f8ab88961e-pod-info" (OuterVolumeSpecName: "pod-info") pod "ce587847-47c2-41de-95d1-c9f8ab88961e" (UID: "ce587847-47c2-41de-95d1-c9f8ab88961e"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.068198 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce587847-47c2-41de-95d1-c9f8ab88961e-config-data" (OuterVolumeSpecName: "config-data") pod "ce587847-47c2-41de-95d1-c9f8ab88961e" (UID: "ce587847-47c2-41de-95d1-c9f8ab88961e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.090843 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce587847-47c2-41de-95d1-c9f8ab88961e-server-conf" (OuterVolumeSpecName: "server-conf") pod "ce587847-47c2-41de-95d1-c9f8ab88961e" (UID: "ce587847-47c2-41de-95d1-c9f8ab88961e"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.113717 4787 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ce587847-47c2-41de-95d1-c9f8ab88961e-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.113764 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqw5k\" (UniqueName: \"kubernetes.io/projected/ce587847-47c2-41de-95d1-c9f8ab88961e-kube-api-access-tqw5k\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.113777 4787 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ce587847-47c2-41de-95d1-c9f8ab88961e-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.113958 4787 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.113972 4787 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ce587847-47c2-41de-95d1-c9f8ab88961e-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.120179 4787 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ce587847-47c2-41de-95d1-c9f8ab88961e-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.120207 4787 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ce587847-47c2-41de-95d1-c9f8ab88961e-pod-info\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.120218 4787 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ce587847-47c2-41de-95d1-c9f8ab88961e-server-conf\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.120229 4787 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ce587847-47c2-41de-95d1-c9f8ab88961e-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.120241 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ce587847-47c2-41de-95d1-c9f8ab88961e-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.136880 4787 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.169541 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce587847-47c2-41de-95d1-c9f8ab88961e-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "ce587847-47c2-41de-95d1-c9f8ab88961e" (UID: "ce587847-47c2-41de-95d1-c9f8ab88961e"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.174461 4787 generic.go:334] "Generic (PLEG): container finished" podID="ce587847-47c2-41de-95d1-c9f8ab88961e" containerID="7fc382e561b660b35dc13b61c6cea44e5887caaef5da418260ae272e6cffb258" exitCode=0 Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.174571 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.175942 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.183355 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ce587847-47c2-41de-95d1-c9f8ab88961e","Type":"ContainerDied","Data":"7fc382e561b660b35dc13b61c6cea44e5887caaef5da418260ae272e6cffb258"} Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.183410 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ce587847-47c2-41de-95d1-c9f8ab88961e","Type":"ContainerDied","Data":"12dc2239ff3366a6d3e4d8d866586456447e31020fbee21027bf57f38d6be118"} Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.183430 4787 scope.go:117] "RemoveContainer" containerID="7fc382e561b660b35dc13b61c6cea44e5887caaef5da418260ae272e6cffb258" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.218369 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.220270 4787 scope.go:117] "RemoveContainer" containerID="e46335d7899902dbb16ab47d03f56851aae7b293726f79ce1a33fe0fda821f6e" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.221805 4787 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ce587847-47c2-41de-95d1-c9f8ab88961e-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.221828 4787 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.229807 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.247930 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 09:55:59 crc kubenswrapper[4787]: E1001 09:55:59.248544 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce587847-47c2-41de-95d1-c9f8ab88961e" containerName="rabbitmq" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.248574 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce587847-47c2-41de-95d1-c9f8ab88961e" containerName="rabbitmq" Oct 01 09:55:59 crc kubenswrapper[4787]: E1001 09:55:59.248593 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce587847-47c2-41de-95d1-c9f8ab88961e" containerName="setup-container" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.248602 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce587847-47c2-41de-95d1-c9f8ab88961e" containerName="setup-container" Oct 01 09:55:59 crc kubenswrapper[4787]: E1001 09:55:59.248644 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e762572f-d5b9-462e-9953-7143f648c9ae" containerName="rabbitmq" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.248653 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e762572f-d5b9-462e-9953-7143f648c9ae" containerName="rabbitmq" Oct 01 09:55:59 crc kubenswrapper[4787]: E1001 09:55:59.248679 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e762572f-d5b9-462e-9953-7143f648c9ae" containerName="setup-container" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.248689 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e762572f-d5b9-462e-9953-7143f648c9ae" containerName="setup-container" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.248971 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce587847-47c2-41de-95d1-c9f8ab88961e" containerName="rabbitmq" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.248990 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="e762572f-d5b9-462e-9953-7143f648c9ae" containerName="rabbitmq" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.250232 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.262573 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-zsk5j" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.262642 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.262795 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.262897 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.264412 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.264506 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.264624 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.273758 4787 scope.go:117] "RemoveContainer" containerID="7fc382e561b660b35dc13b61c6cea44e5887caaef5da418260ae272e6cffb258" Oct 01 09:55:59 crc kubenswrapper[4787]: E1001 09:55:59.282258 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7fc382e561b660b35dc13b61c6cea44e5887caaef5da418260ae272e6cffb258\": container with ID starting with 7fc382e561b660b35dc13b61c6cea44e5887caaef5da418260ae272e6cffb258 not found: ID does not exist" containerID="7fc382e561b660b35dc13b61c6cea44e5887caaef5da418260ae272e6cffb258" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.282327 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fc382e561b660b35dc13b61c6cea44e5887caaef5da418260ae272e6cffb258"} err="failed to get container status \"7fc382e561b660b35dc13b61c6cea44e5887caaef5da418260ae272e6cffb258\": rpc error: code = NotFound desc = could not find container \"7fc382e561b660b35dc13b61c6cea44e5887caaef5da418260ae272e6cffb258\": container with ID starting with 7fc382e561b660b35dc13b61c6cea44e5887caaef5da418260ae272e6cffb258 not found: ID does not exist" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.282371 4787 scope.go:117] "RemoveContainer" containerID="e46335d7899902dbb16ab47d03f56851aae7b293726f79ce1a33fe0fda821f6e" Oct 01 09:55:59 crc kubenswrapper[4787]: E1001 09:55:59.285097 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e46335d7899902dbb16ab47d03f56851aae7b293726f79ce1a33fe0fda821f6e\": container with ID starting with e46335d7899902dbb16ab47d03f56851aae7b293726f79ce1a33fe0fda821f6e not found: ID does not exist" containerID="e46335d7899902dbb16ab47d03f56851aae7b293726f79ce1a33fe0fda821f6e" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.285155 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e46335d7899902dbb16ab47d03f56851aae7b293726f79ce1a33fe0fda821f6e"} err="failed to get container status \"e46335d7899902dbb16ab47d03f56851aae7b293726f79ce1a33fe0fda821f6e\": rpc error: code = NotFound desc = could not find container \"e46335d7899902dbb16ab47d03f56851aae7b293726f79ce1a33fe0fda821f6e\": container with ID starting with e46335d7899902dbb16ab47d03f56851aae7b293726f79ce1a33fe0fda821f6e not found: ID does not exist" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.285218 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.308407 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.315149 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.323145 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.324961 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.329454 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.329627 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-rk8pz" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.329920 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.330100 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.330241 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.330444 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.330609 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.331039 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/eb5022f8-276b-4d6d-917b-be96a1d993bf-server-conf\") pod \"rabbitmq-server-0\" (UID: \"eb5022f8-276b-4d6d-917b-be96a1d993bf\") " pod="openstack/rabbitmq-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.331143 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"eb5022f8-276b-4d6d-917b-be96a1d993bf\") " pod="openstack/rabbitmq-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.331176 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/eb5022f8-276b-4d6d-917b-be96a1d993bf-pod-info\") pod \"rabbitmq-server-0\" (UID: \"eb5022f8-276b-4d6d-917b-be96a1d993bf\") " pod="openstack/rabbitmq-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.331215 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/eb5022f8-276b-4d6d-917b-be96a1d993bf-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"eb5022f8-276b-4d6d-917b-be96a1d993bf\") " pod="openstack/rabbitmq-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.331261 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/eb5022f8-276b-4d6d-917b-be96a1d993bf-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"eb5022f8-276b-4d6d-917b-be96a1d993bf\") " pod="openstack/rabbitmq-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.331322 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/eb5022f8-276b-4d6d-917b-be96a1d993bf-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"eb5022f8-276b-4d6d-917b-be96a1d993bf\") " pod="openstack/rabbitmq-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.331347 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/eb5022f8-276b-4d6d-917b-be96a1d993bf-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"eb5022f8-276b-4d6d-917b-be96a1d993bf\") " pod="openstack/rabbitmq-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.331385 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eb5022f8-276b-4d6d-917b-be96a1d993bf-config-data\") pod \"rabbitmq-server-0\" (UID: \"eb5022f8-276b-4d6d-917b-be96a1d993bf\") " pod="openstack/rabbitmq-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.331403 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/eb5022f8-276b-4d6d-917b-be96a1d993bf-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"eb5022f8-276b-4d6d-917b-be96a1d993bf\") " pod="openstack/rabbitmq-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.331424 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/eb5022f8-276b-4d6d-917b-be96a1d993bf-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"eb5022f8-276b-4d6d-917b-be96a1d993bf\") " pod="openstack/rabbitmq-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.331502 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sm6nl\" (UniqueName: \"kubernetes.io/projected/eb5022f8-276b-4d6d-917b-be96a1d993bf-kube-api-access-sm6nl\") pod \"rabbitmq-server-0\" (UID: \"eb5022f8-276b-4d6d-917b-be96a1d993bf\") " pod="openstack/rabbitmq-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.343645 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.433085 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/eb5022f8-276b-4d6d-917b-be96a1d993bf-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"eb5022f8-276b-4d6d-917b-be96a1d993bf\") " pod="openstack/rabbitmq-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.433157 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d0353483-c978-4d9a-9c34-73ee379325ba-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0353483-c978-4d9a-9c34-73ee379325ba\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.433186 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/eb5022f8-276b-4d6d-917b-be96a1d993bf-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"eb5022f8-276b-4d6d-917b-be96a1d993bf\") " pod="openstack/rabbitmq-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.433206 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d0353483-c978-4d9a-9c34-73ee379325ba-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0353483-c978-4d9a-9c34-73ee379325ba\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.433241 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/eb5022f8-276b-4d6d-917b-be96a1d993bf-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"eb5022f8-276b-4d6d-917b-be96a1d993bf\") " pod="openstack/rabbitmq-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.433255 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/eb5022f8-276b-4d6d-917b-be96a1d993bf-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"eb5022f8-276b-4d6d-917b-be96a1d993bf\") " pod="openstack/rabbitmq-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.433279 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d0353483-c978-4d9a-9c34-73ee379325ba-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0353483-c978-4d9a-9c34-73ee379325ba\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.433303 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eb5022f8-276b-4d6d-917b-be96a1d993bf-config-data\") pod \"rabbitmq-server-0\" (UID: \"eb5022f8-276b-4d6d-917b-be96a1d993bf\") " pod="openstack/rabbitmq-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.433321 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/eb5022f8-276b-4d6d-917b-be96a1d993bf-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"eb5022f8-276b-4d6d-917b-be96a1d993bf\") " pod="openstack/rabbitmq-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.433336 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/eb5022f8-276b-4d6d-917b-be96a1d993bf-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"eb5022f8-276b-4d6d-917b-be96a1d993bf\") " pod="openstack/rabbitmq-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.433364 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d0353483-c978-4d9a-9c34-73ee379325ba-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0353483-c978-4d9a-9c34-73ee379325ba\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.433395 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sm6nl\" (UniqueName: \"kubernetes.io/projected/eb5022f8-276b-4d6d-917b-be96a1d993bf-kube-api-access-sm6nl\") pod \"rabbitmq-server-0\" (UID: \"eb5022f8-276b-4d6d-917b-be96a1d993bf\") " pod="openstack/rabbitmq-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.433415 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d0353483-c978-4d9a-9c34-73ee379325ba-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0353483-c978-4d9a-9c34-73ee379325ba\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.433433 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d0353483-c978-4d9a-9c34-73ee379325ba-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0353483-c978-4d9a-9c34-73ee379325ba\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.433454 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d0353483-c978-4d9a-9c34-73ee379325ba-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0353483-c978-4d9a-9c34-73ee379325ba\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.433473 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d0353483-c978-4d9a-9c34-73ee379325ba-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0353483-c978-4d9a-9c34-73ee379325ba\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.433514 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/eb5022f8-276b-4d6d-917b-be96a1d993bf-server-conf\") pod \"rabbitmq-server-0\" (UID: \"eb5022f8-276b-4d6d-917b-be96a1d993bf\") " pod="openstack/rabbitmq-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.433535 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0353483-c978-4d9a-9c34-73ee379325ba\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.433552 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d0353483-c978-4d9a-9c34-73ee379325ba-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0353483-c978-4d9a-9c34-73ee379325ba\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.433570 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kf2gs\" (UniqueName: \"kubernetes.io/projected/d0353483-c978-4d9a-9c34-73ee379325ba-kube-api-access-kf2gs\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0353483-c978-4d9a-9c34-73ee379325ba\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.433596 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"eb5022f8-276b-4d6d-917b-be96a1d993bf\") " pod="openstack/rabbitmq-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.433630 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/eb5022f8-276b-4d6d-917b-be96a1d993bf-pod-info\") pod \"rabbitmq-server-0\" (UID: \"eb5022f8-276b-4d6d-917b-be96a1d993bf\") " pod="openstack/rabbitmq-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.434535 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/eb5022f8-276b-4d6d-917b-be96a1d993bf-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"eb5022f8-276b-4d6d-917b-be96a1d993bf\") " pod="openstack/rabbitmq-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.434694 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/eb5022f8-276b-4d6d-917b-be96a1d993bf-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"eb5022f8-276b-4d6d-917b-be96a1d993bf\") " pod="openstack/rabbitmq-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.435059 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"eb5022f8-276b-4d6d-917b-be96a1d993bf\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.435439 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/eb5022f8-276b-4d6d-917b-be96a1d993bf-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"eb5022f8-276b-4d6d-917b-be96a1d993bf\") " pod="openstack/rabbitmq-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.435780 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eb5022f8-276b-4d6d-917b-be96a1d993bf-config-data\") pod \"rabbitmq-server-0\" (UID: \"eb5022f8-276b-4d6d-917b-be96a1d993bf\") " pod="openstack/rabbitmq-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.436008 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/eb5022f8-276b-4d6d-917b-be96a1d993bf-server-conf\") pod \"rabbitmq-server-0\" (UID: \"eb5022f8-276b-4d6d-917b-be96a1d993bf\") " pod="openstack/rabbitmq-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.440422 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/eb5022f8-276b-4d6d-917b-be96a1d993bf-pod-info\") pod \"rabbitmq-server-0\" (UID: \"eb5022f8-276b-4d6d-917b-be96a1d993bf\") " pod="openstack/rabbitmq-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.440637 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/eb5022f8-276b-4d6d-917b-be96a1d993bf-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"eb5022f8-276b-4d6d-917b-be96a1d993bf\") " pod="openstack/rabbitmq-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.456523 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/eb5022f8-276b-4d6d-917b-be96a1d993bf-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"eb5022f8-276b-4d6d-917b-be96a1d993bf\") " pod="openstack/rabbitmq-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.470204 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/eb5022f8-276b-4d6d-917b-be96a1d993bf-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"eb5022f8-276b-4d6d-917b-be96a1d993bf\") " pod="openstack/rabbitmq-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.473301 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sm6nl\" (UniqueName: \"kubernetes.io/projected/eb5022f8-276b-4d6d-917b-be96a1d993bf-kube-api-access-sm6nl\") pod \"rabbitmq-server-0\" (UID: \"eb5022f8-276b-4d6d-917b-be96a1d993bf\") " pod="openstack/rabbitmq-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.485551 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"eb5022f8-276b-4d6d-917b-be96a1d993bf\") " pod="openstack/rabbitmq-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.535199 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0353483-c978-4d9a-9c34-73ee379325ba\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.535241 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d0353483-c978-4d9a-9c34-73ee379325ba-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0353483-c978-4d9a-9c34-73ee379325ba\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.535266 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kf2gs\" (UniqueName: \"kubernetes.io/projected/d0353483-c978-4d9a-9c34-73ee379325ba-kube-api-access-kf2gs\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0353483-c978-4d9a-9c34-73ee379325ba\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.535352 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d0353483-c978-4d9a-9c34-73ee379325ba-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0353483-c978-4d9a-9c34-73ee379325ba\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.535371 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d0353483-c978-4d9a-9c34-73ee379325ba-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0353483-c978-4d9a-9c34-73ee379325ba\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.535436 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d0353483-c978-4d9a-9c34-73ee379325ba-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0353483-c978-4d9a-9c34-73ee379325ba\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.535482 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d0353483-c978-4d9a-9c34-73ee379325ba-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0353483-c978-4d9a-9c34-73ee379325ba\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.535542 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d0353483-c978-4d9a-9c34-73ee379325ba-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0353483-c978-4d9a-9c34-73ee379325ba\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.535576 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d0353483-c978-4d9a-9c34-73ee379325ba-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0353483-c978-4d9a-9c34-73ee379325ba\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.535605 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d0353483-c978-4d9a-9c34-73ee379325ba-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0353483-c978-4d9a-9c34-73ee379325ba\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.535624 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d0353483-c978-4d9a-9c34-73ee379325ba-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0353483-c978-4d9a-9c34-73ee379325ba\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.536103 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d0353483-c978-4d9a-9c34-73ee379325ba-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0353483-c978-4d9a-9c34-73ee379325ba\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.535532 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0353483-c978-4d9a-9c34-73ee379325ba\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.536766 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d0353483-c978-4d9a-9c34-73ee379325ba-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0353483-c978-4d9a-9c34-73ee379325ba\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.537238 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d0353483-c978-4d9a-9c34-73ee379325ba-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0353483-c978-4d9a-9c34-73ee379325ba\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.537818 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d0353483-c978-4d9a-9c34-73ee379325ba-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0353483-c978-4d9a-9c34-73ee379325ba\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.538369 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d0353483-c978-4d9a-9c34-73ee379325ba-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0353483-c978-4d9a-9c34-73ee379325ba\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.540712 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d0353483-c978-4d9a-9c34-73ee379325ba-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0353483-c978-4d9a-9c34-73ee379325ba\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.541238 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d0353483-c978-4d9a-9c34-73ee379325ba-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0353483-c978-4d9a-9c34-73ee379325ba\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.542403 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d0353483-c978-4d9a-9c34-73ee379325ba-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0353483-c978-4d9a-9c34-73ee379325ba\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.542430 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d0353483-c978-4d9a-9c34-73ee379325ba-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0353483-c978-4d9a-9c34-73ee379325ba\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.554524 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kf2gs\" (UniqueName: \"kubernetes.io/projected/d0353483-c978-4d9a-9c34-73ee379325ba-kube-api-access-kf2gs\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0353483-c978-4d9a-9c34-73ee379325ba\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.572481 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d0353483-c978-4d9a-9c34-73ee379325ba\") " pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.583769 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 01 09:55:59 crc kubenswrapper[4787]: I1001 09:55:59.644815 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:56:00 crc kubenswrapper[4787]: I1001 09:56:00.540222 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce587847-47c2-41de-95d1-c9f8ab88961e" path="/var/lib/kubelet/pods/ce587847-47c2-41de-95d1-c9f8ab88961e/volumes" Oct 01 09:56:00 crc kubenswrapper[4787]: I1001 09:56:00.541598 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e762572f-d5b9-462e-9953-7143f648c9ae" path="/var/lib/kubelet/pods/e762572f-d5b9-462e-9953-7143f648c9ae/volumes" Oct 01 09:56:00 crc kubenswrapper[4787]: I1001 09:56:00.654935 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 01 09:56:00 crc kubenswrapper[4787]: W1001 09:56:00.763438 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd0353483_c978_4d9a_9c34_73ee379325ba.slice/crio-577af882868f48ed51220b44e0e8af4753e7539ee5c941a240a5ebdc6faeeede WatchSource:0}: Error finding container 577af882868f48ed51220b44e0e8af4753e7539ee5c941a240a5ebdc6faeeede: Status 404 returned error can't find the container with id 577af882868f48ed51220b44e0e8af4753e7539ee5c941a240a5ebdc6faeeede Oct 01 09:56:00 crc kubenswrapper[4787]: I1001 09:56:00.764365 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.070032 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-766f8885ff-qm2wv"] Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.075523 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-766f8885ff-qm2wv" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.084054 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-766f8885ff-qm2wv"] Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.084836 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.167064 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-dns-svc\") pod \"dnsmasq-dns-766f8885ff-qm2wv\" (UID: \"ce6c895d-bf1f-4182-a6d4-02bb1c964cae\") " pod="openstack/dnsmasq-dns-766f8885ff-qm2wv" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.167135 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-ovsdbserver-sb\") pod \"dnsmasq-dns-766f8885ff-qm2wv\" (UID: \"ce6c895d-bf1f-4182-a6d4-02bb1c964cae\") " pod="openstack/dnsmasq-dns-766f8885ff-qm2wv" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.167168 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-ovsdbserver-nb\") pod \"dnsmasq-dns-766f8885ff-qm2wv\" (UID: \"ce6c895d-bf1f-4182-a6d4-02bb1c964cae\") " pod="openstack/dnsmasq-dns-766f8885ff-qm2wv" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.167195 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-openstack-edpm-ipam\") pod \"dnsmasq-dns-766f8885ff-qm2wv\" (UID: \"ce6c895d-bf1f-4182-a6d4-02bb1c964cae\") " pod="openstack/dnsmasq-dns-766f8885ff-qm2wv" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.167240 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-config\") pod \"dnsmasq-dns-766f8885ff-qm2wv\" (UID: \"ce6c895d-bf1f-4182-a6d4-02bb1c964cae\") " pod="openstack/dnsmasq-dns-766f8885ff-qm2wv" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.167284 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7zz2\" (UniqueName: \"kubernetes.io/projected/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-kube-api-access-w7zz2\") pod \"dnsmasq-dns-766f8885ff-qm2wv\" (UID: \"ce6c895d-bf1f-4182-a6d4-02bb1c964cae\") " pod="openstack/dnsmasq-dns-766f8885ff-qm2wv" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.167339 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-dns-swift-storage-0\") pod \"dnsmasq-dns-766f8885ff-qm2wv\" (UID: \"ce6c895d-bf1f-4182-a6d4-02bb1c964cae\") " pod="openstack/dnsmasq-dns-766f8885ff-qm2wv" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.202023 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"eb5022f8-276b-4d6d-917b-be96a1d993bf","Type":"ContainerStarted","Data":"3210746e32acdc661daa0f2044da0144db511d7338ff9657472194ee65a3014a"} Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.203817 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d0353483-c978-4d9a-9c34-73ee379325ba","Type":"ContainerStarted","Data":"577af882868f48ed51220b44e0e8af4753e7539ee5c941a240a5ebdc6faeeede"} Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.207503 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-766f8885ff-qm2wv"] Oct 01 09:56:01 crc kubenswrapper[4787]: E1001 09:56:01.208336 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc dns-swift-storage-0 kube-api-access-w7zz2 openstack-edpm-ipam ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-766f8885ff-qm2wv" podUID="ce6c895d-bf1f-4182-a6d4-02bb1c964cae" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.239332 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6966dc867f-cvjjh"] Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.241229 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6966dc867f-cvjjh" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.257439 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6966dc867f-cvjjh"] Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.268491 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqt68\" (UniqueName: \"kubernetes.io/projected/fc33ac30-efd1-438d-9ab5-d700dfd27efe-kube-api-access-cqt68\") pod \"dnsmasq-dns-6966dc867f-cvjjh\" (UID: \"fc33ac30-efd1-438d-9ab5-d700dfd27efe\") " pod="openstack/dnsmasq-dns-6966dc867f-cvjjh" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.268543 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/fc33ac30-efd1-438d-9ab5-d700dfd27efe-openstack-edpm-ipam\") pod \"dnsmasq-dns-6966dc867f-cvjjh\" (UID: \"fc33ac30-efd1-438d-9ab5-d700dfd27efe\") " pod="openstack/dnsmasq-dns-6966dc867f-cvjjh" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.268559 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fc33ac30-efd1-438d-9ab5-d700dfd27efe-dns-swift-storage-0\") pod \"dnsmasq-dns-6966dc867f-cvjjh\" (UID: \"fc33ac30-efd1-438d-9ab5-d700dfd27efe\") " pod="openstack/dnsmasq-dns-6966dc867f-cvjjh" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.268627 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-dns-swift-storage-0\") pod \"dnsmasq-dns-766f8885ff-qm2wv\" (UID: \"ce6c895d-bf1f-4182-a6d4-02bb1c964cae\") " pod="openstack/dnsmasq-dns-766f8885ff-qm2wv" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.268679 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc33ac30-efd1-438d-9ab5-d700dfd27efe-dns-svc\") pod \"dnsmasq-dns-6966dc867f-cvjjh\" (UID: \"fc33ac30-efd1-438d-9ab5-d700dfd27efe\") " pod="openstack/dnsmasq-dns-6966dc867f-cvjjh" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.268702 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc33ac30-efd1-438d-9ab5-d700dfd27efe-config\") pod \"dnsmasq-dns-6966dc867f-cvjjh\" (UID: \"fc33ac30-efd1-438d-9ab5-d700dfd27efe\") " pod="openstack/dnsmasq-dns-6966dc867f-cvjjh" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.268754 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-dns-svc\") pod \"dnsmasq-dns-766f8885ff-qm2wv\" (UID: \"ce6c895d-bf1f-4182-a6d4-02bb1c964cae\") " pod="openstack/dnsmasq-dns-766f8885ff-qm2wv" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.268780 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-ovsdbserver-sb\") pod \"dnsmasq-dns-766f8885ff-qm2wv\" (UID: \"ce6c895d-bf1f-4182-a6d4-02bb1c964cae\") " pod="openstack/dnsmasq-dns-766f8885ff-qm2wv" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.268799 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-ovsdbserver-nb\") pod \"dnsmasq-dns-766f8885ff-qm2wv\" (UID: \"ce6c895d-bf1f-4182-a6d4-02bb1c964cae\") " pod="openstack/dnsmasq-dns-766f8885ff-qm2wv" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.268826 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fc33ac30-efd1-438d-9ab5-d700dfd27efe-ovsdbserver-sb\") pod \"dnsmasq-dns-6966dc867f-cvjjh\" (UID: \"fc33ac30-efd1-438d-9ab5-d700dfd27efe\") " pod="openstack/dnsmasq-dns-6966dc867f-cvjjh" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.268845 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fc33ac30-efd1-438d-9ab5-d700dfd27efe-ovsdbserver-nb\") pod \"dnsmasq-dns-6966dc867f-cvjjh\" (UID: \"fc33ac30-efd1-438d-9ab5-d700dfd27efe\") " pod="openstack/dnsmasq-dns-6966dc867f-cvjjh" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.268867 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-openstack-edpm-ipam\") pod \"dnsmasq-dns-766f8885ff-qm2wv\" (UID: \"ce6c895d-bf1f-4182-a6d4-02bb1c964cae\") " pod="openstack/dnsmasq-dns-766f8885ff-qm2wv" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.268936 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-config\") pod \"dnsmasq-dns-766f8885ff-qm2wv\" (UID: \"ce6c895d-bf1f-4182-a6d4-02bb1c964cae\") " pod="openstack/dnsmasq-dns-766f8885ff-qm2wv" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.269214 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7zz2\" (UniqueName: \"kubernetes.io/projected/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-kube-api-access-w7zz2\") pod \"dnsmasq-dns-766f8885ff-qm2wv\" (UID: \"ce6c895d-bf1f-4182-a6d4-02bb1c964cae\") " pod="openstack/dnsmasq-dns-766f8885ff-qm2wv" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.269811 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-openstack-edpm-ipam\") pod \"dnsmasq-dns-766f8885ff-qm2wv\" (UID: \"ce6c895d-bf1f-4182-a6d4-02bb1c964cae\") " pod="openstack/dnsmasq-dns-766f8885ff-qm2wv" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.269842 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-ovsdbserver-sb\") pod \"dnsmasq-dns-766f8885ff-qm2wv\" (UID: \"ce6c895d-bf1f-4182-a6d4-02bb1c964cae\") " pod="openstack/dnsmasq-dns-766f8885ff-qm2wv" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.269899 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-dns-swift-storage-0\") pod \"dnsmasq-dns-766f8885ff-qm2wv\" (UID: \"ce6c895d-bf1f-4182-a6d4-02bb1c964cae\") " pod="openstack/dnsmasq-dns-766f8885ff-qm2wv" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.270402 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-config\") pod \"dnsmasq-dns-766f8885ff-qm2wv\" (UID: \"ce6c895d-bf1f-4182-a6d4-02bb1c964cae\") " pod="openstack/dnsmasq-dns-766f8885ff-qm2wv" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.270448 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-ovsdbserver-nb\") pod \"dnsmasq-dns-766f8885ff-qm2wv\" (UID: \"ce6c895d-bf1f-4182-a6d4-02bb1c964cae\") " pod="openstack/dnsmasq-dns-766f8885ff-qm2wv" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.270743 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-dns-svc\") pod \"dnsmasq-dns-766f8885ff-qm2wv\" (UID: \"ce6c895d-bf1f-4182-a6d4-02bb1c964cae\") " pod="openstack/dnsmasq-dns-766f8885ff-qm2wv" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.293617 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7zz2\" (UniqueName: \"kubernetes.io/projected/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-kube-api-access-w7zz2\") pod \"dnsmasq-dns-766f8885ff-qm2wv\" (UID: \"ce6c895d-bf1f-4182-a6d4-02bb1c964cae\") " pod="openstack/dnsmasq-dns-766f8885ff-qm2wv" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.371318 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqt68\" (UniqueName: \"kubernetes.io/projected/fc33ac30-efd1-438d-9ab5-d700dfd27efe-kube-api-access-cqt68\") pod \"dnsmasq-dns-6966dc867f-cvjjh\" (UID: \"fc33ac30-efd1-438d-9ab5-d700dfd27efe\") " pod="openstack/dnsmasq-dns-6966dc867f-cvjjh" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.371405 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/fc33ac30-efd1-438d-9ab5-d700dfd27efe-openstack-edpm-ipam\") pod \"dnsmasq-dns-6966dc867f-cvjjh\" (UID: \"fc33ac30-efd1-438d-9ab5-d700dfd27efe\") " pod="openstack/dnsmasq-dns-6966dc867f-cvjjh" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.371447 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fc33ac30-efd1-438d-9ab5-d700dfd27efe-dns-swift-storage-0\") pod \"dnsmasq-dns-6966dc867f-cvjjh\" (UID: \"fc33ac30-efd1-438d-9ab5-d700dfd27efe\") " pod="openstack/dnsmasq-dns-6966dc867f-cvjjh" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.371527 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc33ac30-efd1-438d-9ab5-d700dfd27efe-dns-svc\") pod \"dnsmasq-dns-6966dc867f-cvjjh\" (UID: \"fc33ac30-efd1-438d-9ab5-d700dfd27efe\") " pod="openstack/dnsmasq-dns-6966dc867f-cvjjh" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.371572 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc33ac30-efd1-438d-9ab5-d700dfd27efe-config\") pod \"dnsmasq-dns-6966dc867f-cvjjh\" (UID: \"fc33ac30-efd1-438d-9ab5-d700dfd27efe\") " pod="openstack/dnsmasq-dns-6966dc867f-cvjjh" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.371670 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fc33ac30-efd1-438d-9ab5-d700dfd27efe-ovsdbserver-sb\") pod \"dnsmasq-dns-6966dc867f-cvjjh\" (UID: \"fc33ac30-efd1-438d-9ab5-d700dfd27efe\") " pod="openstack/dnsmasq-dns-6966dc867f-cvjjh" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.371702 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fc33ac30-efd1-438d-9ab5-d700dfd27efe-ovsdbserver-nb\") pod \"dnsmasq-dns-6966dc867f-cvjjh\" (UID: \"fc33ac30-efd1-438d-9ab5-d700dfd27efe\") " pod="openstack/dnsmasq-dns-6966dc867f-cvjjh" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.373291 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fc33ac30-efd1-438d-9ab5-d700dfd27efe-ovsdbserver-nb\") pod \"dnsmasq-dns-6966dc867f-cvjjh\" (UID: \"fc33ac30-efd1-438d-9ab5-d700dfd27efe\") " pod="openstack/dnsmasq-dns-6966dc867f-cvjjh" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.374861 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc33ac30-efd1-438d-9ab5-d700dfd27efe-dns-svc\") pod \"dnsmasq-dns-6966dc867f-cvjjh\" (UID: \"fc33ac30-efd1-438d-9ab5-d700dfd27efe\") " pod="openstack/dnsmasq-dns-6966dc867f-cvjjh" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.375712 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc33ac30-efd1-438d-9ab5-d700dfd27efe-config\") pod \"dnsmasq-dns-6966dc867f-cvjjh\" (UID: \"fc33ac30-efd1-438d-9ab5-d700dfd27efe\") " pod="openstack/dnsmasq-dns-6966dc867f-cvjjh" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.376165 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fc33ac30-efd1-438d-9ab5-d700dfd27efe-ovsdbserver-sb\") pod \"dnsmasq-dns-6966dc867f-cvjjh\" (UID: \"fc33ac30-efd1-438d-9ab5-d700dfd27efe\") " pod="openstack/dnsmasq-dns-6966dc867f-cvjjh" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.377387 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fc33ac30-efd1-438d-9ab5-d700dfd27efe-dns-swift-storage-0\") pod \"dnsmasq-dns-6966dc867f-cvjjh\" (UID: \"fc33ac30-efd1-438d-9ab5-d700dfd27efe\") " pod="openstack/dnsmasq-dns-6966dc867f-cvjjh" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.377915 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/fc33ac30-efd1-438d-9ab5-d700dfd27efe-openstack-edpm-ipam\") pod \"dnsmasq-dns-6966dc867f-cvjjh\" (UID: \"fc33ac30-efd1-438d-9ab5-d700dfd27efe\") " pod="openstack/dnsmasq-dns-6966dc867f-cvjjh" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.390899 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqt68\" (UniqueName: \"kubernetes.io/projected/fc33ac30-efd1-438d-9ab5-d700dfd27efe-kube-api-access-cqt68\") pod \"dnsmasq-dns-6966dc867f-cvjjh\" (UID: \"fc33ac30-efd1-438d-9ab5-d700dfd27efe\") " pod="openstack/dnsmasq-dns-6966dc867f-cvjjh" Oct 01 09:56:01 crc kubenswrapper[4787]: I1001 09:56:01.557578 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6966dc867f-cvjjh" Oct 01 09:56:02 crc kubenswrapper[4787]: I1001 09:56:02.240252 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-766f8885ff-qm2wv" Oct 01 09:56:02 crc kubenswrapper[4787]: I1001 09:56:02.273365 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-766f8885ff-qm2wv" Oct 01 09:56:02 crc kubenswrapper[4787]: I1001 09:56:02.287357 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6966dc867f-cvjjh"] Oct 01 09:56:02 crc kubenswrapper[4787]: I1001 09:56:02.289626 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-dns-swift-storage-0\") pod \"ce6c895d-bf1f-4182-a6d4-02bb1c964cae\" (UID: \"ce6c895d-bf1f-4182-a6d4-02bb1c964cae\") " Oct 01 09:56:02 crc kubenswrapper[4787]: I1001 09:56:02.289806 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-openstack-edpm-ipam\") pod \"ce6c895d-bf1f-4182-a6d4-02bb1c964cae\" (UID: \"ce6c895d-bf1f-4182-a6d4-02bb1c964cae\") " Oct 01 09:56:02 crc kubenswrapper[4787]: I1001 09:56:02.289941 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-config\") pod \"ce6c895d-bf1f-4182-a6d4-02bb1c964cae\" (UID: \"ce6c895d-bf1f-4182-a6d4-02bb1c964cae\") " Oct 01 09:56:02 crc kubenswrapper[4787]: I1001 09:56:02.290012 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-ovsdbserver-nb\") pod \"ce6c895d-bf1f-4182-a6d4-02bb1c964cae\" (UID: \"ce6c895d-bf1f-4182-a6d4-02bb1c964cae\") " Oct 01 09:56:02 crc kubenswrapper[4787]: I1001 09:56:02.290175 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7zz2\" (UniqueName: \"kubernetes.io/projected/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-kube-api-access-w7zz2\") pod \"ce6c895d-bf1f-4182-a6d4-02bb1c964cae\" (UID: \"ce6c895d-bf1f-4182-a6d4-02bb1c964cae\") " Oct 01 09:56:02 crc kubenswrapper[4787]: I1001 09:56:02.290277 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-dns-svc\") pod \"ce6c895d-bf1f-4182-a6d4-02bb1c964cae\" (UID: \"ce6c895d-bf1f-4182-a6d4-02bb1c964cae\") " Oct 01 09:56:02 crc kubenswrapper[4787]: I1001 09:56:02.290340 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "ce6c895d-bf1f-4182-a6d4-02bb1c964cae" (UID: "ce6c895d-bf1f-4182-a6d4-02bb1c964cae"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:56:02 crc kubenswrapper[4787]: I1001 09:56:02.290362 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-config" (OuterVolumeSpecName: "config") pod "ce6c895d-bf1f-4182-a6d4-02bb1c964cae" (UID: "ce6c895d-bf1f-4182-a6d4-02bb1c964cae"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:56:02 crc kubenswrapper[4787]: I1001 09:56:02.290536 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-ovsdbserver-sb\") pod \"ce6c895d-bf1f-4182-a6d4-02bb1c964cae\" (UID: \"ce6c895d-bf1f-4182-a6d4-02bb1c964cae\") " Oct 01 09:56:02 crc kubenswrapper[4787]: I1001 09:56:02.290682 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ce6c895d-bf1f-4182-a6d4-02bb1c964cae" (UID: "ce6c895d-bf1f-4182-a6d4-02bb1c964cae"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:56:02 crc kubenswrapper[4787]: I1001 09:56:02.290998 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ce6c895d-bf1f-4182-a6d4-02bb1c964cae" (UID: "ce6c895d-bf1f-4182-a6d4-02bb1c964cae"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:56:02 crc kubenswrapper[4787]: I1001 09:56:02.291129 4787 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 01 09:56:02 crc kubenswrapper[4787]: I1001 09:56:02.291190 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 09:56:02 crc kubenswrapper[4787]: I1001 09:56:02.291240 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:56:02 crc kubenswrapper[4787]: I1001 09:56:02.291455 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ce6c895d-bf1f-4182-a6d4-02bb1c964cae" (UID: "ce6c895d-bf1f-4182-a6d4-02bb1c964cae"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:56:02 crc kubenswrapper[4787]: I1001 09:56:02.291600 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ce6c895d-bf1f-4182-a6d4-02bb1c964cae" (UID: "ce6c895d-bf1f-4182-a6d4-02bb1c964cae"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:56:02 crc kubenswrapper[4787]: I1001 09:56:02.296117 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-kube-api-access-w7zz2" (OuterVolumeSpecName: "kube-api-access-w7zz2") pod "ce6c895d-bf1f-4182-a6d4-02bb1c964cae" (UID: "ce6c895d-bf1f-4182-a6d4-02bb1c964cae"). InnerVolumeSpecName "kube-api-access-w7zz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:56:02 crc kubenswrapper[4787]: W1001 09:56:02.299209 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfc33ac30_efd1_438d_9ab5_d700dfd27efe.slice/crio-3cfbe7c23d02df37c6d31ec668d55efe0fc505a06c4cc590d581c39a5a2ee947 WatchSource:0}: Error finding container 3cfbe7c23d02df37c6d31ec668d55efe0fc505a06c4cc590d581c39a5a2ee947: Status 404 returned error can't find the container with id 3cfbe7c23d02df37c6d31ec668d55efe0fc505a06c4cc590d581c39a5a2ee947 Oct 01 09:56:02 crc kubenswrapper[4787]: I1001 09:56:02.393578 4787 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 09:56:02 crc kubenswrapper[4787]: I1001 09:56:02.393798 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 09:56:02 crc kubenswrapper[4787]: I1001 09:56:02.393809 4787 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 01 09:56:02 crc kubenswrapper[4787]: I1001 09:56:02.393818 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7zz2\" (UniqueName: \"kubernetes.io/projected/ce6c895d-bf1f-4182-a6d4-02bb1c964cae-kube-api-access-w7zz2\") on node \"crc\" DevicePath \"\"" Oct 01 09:56:03 crc kubenswrapper[4787]: I1001 09:56:03.254741 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"eb5022f8-276b-4d6d-917b-be96a1d993bf","Type":"ContainerStarted","Data":"8043fd4f5d9a066ef9a0029929fd72592afd40dbb50a23407e5a155c8f2364aa"} Oct 01 09:56:03 crc kubenswrapper[4787]: I1001 09:56:03.257468 4787 generic.go:334] "Generic (PLEG): container finished" podID="fc33ac30-efd1-438d-9ab5-d700dfd27efe" containerID="89640bbe33894d44d8ee7ee15156c57591172e0fdc4731e0f69eb21350372923" exitCode=0 Oct 01 09:56:03 crc kubenswrapper[4787]: I1001 09:56:03.257584 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-766f8885ff-qm2wv" Oct 01 09:56:03 crc kubenswrapper[4787]: I1001 09:56:03.258922 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6966dc867f-cvjjh" event={"ID":"fc33ac30-efd1-438d-9ab5-d700dfd27efe","Type":"ContainerDied","Data":"89640bbe33894d44d8ee7ee15156c57591172e0fdc4731e0f69eb21350372923"} Oct 01 09:56:03 crc kubenswrapper[4787]: I1001 09:56:03.258953 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6966dc867f-cvjjh" event={"ID":"fc33ac30-efd1-438d-9ab5-d700dfd27efe","Type":"ContainerStarted","Data":"3cfbe7c23d02df37c6d31ec668d55efe0fc505a06c4cc590d581c39a5a2ee947"} Oct 01 09:56:03 crc kubenswrapper[4787]: I1001 09:56:03.490269 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-766f8885ff-qm2wv"] Oct 01 09:56:03 crc kubenswrapper[4787]: I1001 09:56:03.498298 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-766f8885ff-qm2wv"] Oct 01 09:56:04 crc kubenswrapper[4787]: I1001 09:56:04.271549 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6966dc867f-cvjjh" event={"ID":"fc33ac30-efd1-438d-9ab5-d700dfd27efe","Type":"ContainerStarted","Data":"c6fcdbbc52370ef5dc80e7b3484304cbd5011420dee635ec04dbcc17f629a731"} Oct 01 09:56:04 crc kubenswrapper[4787]: I1001 09:56:04.272167 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6966dc867f-cvjjh" Oct 01 09:56:04 crc kubenswrapper[4787]: I1001 09:56:04.273921 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d0353483-c978-4d9a-9c34-73ee379325ba","Type":"ContainerStarted","Data":"03af902c2fd8ec77a8dd910f7e1f4b6f157e978fef61cc57265b74c68df142b4"} Oct 01 09:56:04 crc kubenswrapper[4787]: I1001 09:56:04.310320 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6966dc867f-cvjjh" podStartSLOduration=3.310280487 podStartE2EDuration="3.310280487s" podCreationTimestamp="2025-10-01 09:56:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:56:04.294278881 +0000 UTC m=+1196.409423098" watchObservedRunningTime="2025-10-01 09:56:04.310280487 +0000 UTC m=+1196.425424694" Oct 01 09:56:04 crc kubenswrapper[4787]: I1001 09:56:04.537430 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce6c895d-bf1f-4182-a6d4-02bb1c964cae" path="/var/lib/kubelet/pods/ce6c895d-bf1f-4182-a6d4-02bb1c964cae/volumes" Oct 01 09:56:11 crc kubenswrapper[4787]: I1001 09:56:11.559290 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6966dc867f-cvjjh" Oct 01 09:56:11 crc kubenswrapper[4787]: I1001 09:56:11.622325 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68b59d98cf-xk49b"] Oct 01 09:56:11 crc kubenswrapper[4787]: I1001 09:56:11.622575 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-68b59d98cf-xk49b" podUID="ed8acab0-e099-4b65-9883-5ea45cc4599e" containerName="dnsmasq-dns" containerID="cri-o://3bf412fd2bd6296d4602149f7c73e11ccd49430d8c5803653fcadeab6637bfa5" gracePeriod=10 Oct 01 09:56:12 crc kubenswrapper[4787]: I1001 09:56:12.111793 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68b59d98cf-xk49b" Oct 01 09:56:12 crc kubenswrapper[4787]: I1001 09:56:12.226193 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed8acab0-e099-4b65-9883-5ea45cc4599e-dns-svc\") pod \"ed8acab0-e099-4b65-9883-5ea45cc4599e\" (UID: \"ed8acab0-e099-4b65-9883-5ea45cc4599e\") " Oct 01 09:56:12 crc kubenswrapper[4787]: I1001 09:56:12.226293 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvk7q\" (UniqueName: \"kubernetes.io/projected/ed8acab0-e099-4b65-9883-5ea45cc4599e-kube-api-access-dvk7q\") pod \"ed8acab0-e099-4b65-9883-5ea45cc4599e\" (UID: \"ed8acab0-e099-4b65-9883-5ea45cc4599e\") " Oct 01 09:56:12 crc kubenswrapper[4787]: I1001 09:56:12.226336 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed8acab0-e099-4b65-9883-5ea45cc4599e-ovsdbserver-nb\") pod \"ed8acab0-e099-4b65-9883-5ea45cc4599e\" (UID: \"ed8acab0-e099-4b65-9883-5ea45cc4599e\") " Oct 01 09:56:12 crc kubenswrapper[4787]: I1001 09:56:12.226398 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ed8acab0-e099-4b65-9883-5ea45cc4599e-dns-swift-storage-0\") pod \"ed8acab0-e099-4b65-9883-5ea45cc4599e\" (UID: \"ed8acab0-e099-4b65-9883-5ea45cc4599e\") " Oct 01 09:56:12 crc kubenswrapper[4787]: I1001 09:56:12.226442 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed8acab0-e099-4b65-9883-5ea45cc4599e-config\") pod \"ed8acab0-e099-4b65-9883-5ea45cc4599e\" (UID: \"ed8acab0-e099-4b65-9883-5ea45cc4599e\") " Oct 01 09:56:12 crc kubenswrapper[4787]: I1001 09:56:12.226499 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed8acab0-e099-4b65-9883-5ea45cc4599e-ovsdbserver-sb\") pod \"ed8acab0-e099-4b65-9883-5ea45cc4599e\" (UID: \"ed8acab0-e099-4b65-9883-5ea45cc4599e\") " Oct 01 09:56:12 crc kubenswrapper[4787]: I1001 09:56:12.232695 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed8acab0-e099-4b65-9883-5ea45cc4599e-kube-api-access-dvk7q" (OuterVolumeSpecName: "kube-api-access-dvk7q") pod "ed8acab0-e099-4b65-9883-5ea45cc4599e" (UID: "ed8acab0-e099-4b65-9883-5ea45cc4599e"). InnerVolumeSpecName "kube-api-access-dvk7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:56:12 crc kubenswrapper[4787]: I1001 09:56:12.280766 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed8acab0-e099-4b65-9883-5ea45cc4599e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ed8acab0-e099-4b65-9883-5ea45cc4599e" (UID: "ed8acab0-e099-4b65-9883-5ea45cc4599e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:56:12 crc kubenswrapper[4787]: I1001 09:56:12.280790 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed8acab0-e099-4b65-9883-5ea45cc4599e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ed8acab0-e099-4b65-9883-5ea45cc4599e" (UID: "ed8acab0-e099-4b65-9883-5ea45cc4599e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:56:12 crc kubenswrapper[4787]: I1001 09:56:12.292498 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed8acab0-e099-4b65-9883-5ea45cc4599e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ed8acab0-e099-4b65-9883-5ea45cc4599e" (UID: "ed8acab0-e099-4b65-9883-5ea45cc4599e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:56:12 crc kubenswrapper[4787]: I1001 09:56:12.301251 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed8acab0-e099-4b65-9883-5ea45cc4599e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ed8acab0-e099-4b65-9883-5ea45cc4599e" (UID: "ed8acab0-e099-4b65-9883-5ea45cc4599e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:56:12 crc kubenswrapper[4787]: I1001 09:56:12.302508 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed8acab0-e099-4b65-9883-5ea45cc4599e-config" (OuterVolumeSpecName: "config") pod "ed8acab0-e099-4b65-9883-5ea45cc4599e" (UID: "ed8acab0-e099-4b65-9883-5ea45cc4599e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 09:56:12 crc kubenswrapper[4787]: I1001 09:56:12.328725 4787 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ed8acab0-e099-4b65-9883-5ea45cc4599e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 01 09:56:12 crc kubenswrapper[4787]: I1001 09:56:12.328768 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed8acab0-e099-4b65-9883-5ea45cc4599e-config\") on node \"crc\" DevicePath \"\"" Oct 01 09:56:12 crc kubenswrapper[4787]: I1001 09:56:12.328793 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed8acab0-e099-4b65-9883-5ea45cc4599e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 01 09:56:12 crc kubenswrapper[4787]: I1001 09:56:12.328812 4787 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed8acab0-e099-4b65-9883-5ea45cc4599e-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 01 09:56:12 crc kubenswrapper[4787]: I1001 09:56:12.328831 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvk7q\" (UniqueName: \"kubernetes.io/projected/ed8acab0-e099-4b65-9883-5ea45cc4599e-kube-api-access-dvk7q\") on node \"crc\" DevicePath \"\"" Oct 01 09:56:12 crc kubenswrapper[4787]: I1001 09:56:12.328850 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed8acab0-e099-4b65-9883-5ea45cc4599e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 01 09:56:12 crc kubenswrapper[4787]: I1001 09:56:12.364326 4787 generic.go:334] "Generic (PLEG): container finished" podID="ed8acab0-e099-4b65-9883-5ea45cc4599e" containerID="3bf412fd2bd6296d4602149f7c73e11ccd49430d8c5803653fcadeab6637bfa5" exitCode=0 Oct 01 09:56:12 crc kubenswrapper[4787]: I1001 09:56:12.364386 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68b59d98cf-xk49b" event={"ID":"ed8acab0-e099-4b65-9883-5ea45cc4599e","Type":"ContainerDied","Data":"3bf412fd2bd6296d4602149f7c73e11ccd49430d8c5803653fcadeab6637bfa5"} Oct 01 09:56:12 crc kubenswrapper[4787]: I1001 09:56:12.364424 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68b59d98cf-xk49b" event={"ID":"ed8acab0-e099-4b65-9883-5ea45cc4599e","Type":"ContainerDied","Data":"6826e6bae9ecd4cbe77a334450c216d02fcc72a1432592f33335bd0a990ce6ec"} Oct 01 09:56:12 crc kubenswrapper[4787]: I1001 09:56:12.364455 4787 scope.go:117] "RemoveContainer" containerID="3bf412fd2bd6296d4602149f7c73e11ccd49430d8c5803653fcadeab6637bfa5" Oct 01 09:56:12 crc kubenswrapper[4787]: I1001 09:56:12.364626 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68b59d98cf-xk49b" Oct 01 09:56:12 crc kubenswrapper[4787]: I1001 09:56:12.401835 4787 scope.go:117] "RemoveContainer" containerID="f1bc5cf1281c1f641d7c326c41bd88fcd42be40d931224664f09b2e9338df0cb" Oct 01 09:56:12 crc kubenswrapper[4787]: I1001 09:56:12.427188 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68b59d98cf-xk49b"] Oct 01 09:56:12 crc kubenswrapper[4787]: I1001 09:56:12.431220 4787 scope.go:117] "RemoveContainer" containerID="3bf412fd2bd6296d4602149f7c73e11ccd49430d8c5803653fcadeab6637bfa5" Oct 01 09:56:12 crc kubenswrapper[4787]: E1001 09:56:12.432996 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3bf412fd2bd6296d4602149f7c73e11ccd49430d8c5803653fcadeab6637bfa5\": container with ID starting with 3bf412fd2bd6296d4602149f7c73e11ccd49430d8c5803653fcadeab6637bfa5 not found: ID does not exist" containerID="3bf412fd2bd6296d4602149f7c73e11ccd49430d8c5803653fcadeab6637bfa5" Oct 01 09:56:12 crc kubenswrapper[4787]: I1001 09:56:12.433086 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bf412fd2bd6296d4602149f7c73e11ccd49430d8c5803653fcadeab6637bfa5"} err="failed to get container status \"3bf412fd2bd6296d4602149f7c73e11ccd49430d8c5803653fcadeab6637bfa5\": rpc error: code = NotFound desc = could not find container \"3bf412fd2bd6296d4602149f7c73e11ccd49430d8c5803653fcadeab6637bfa5\": container with ID starting with 3bf412fd2bd6296d4602149f7c73e11ccd49430d8c5803653fcadeab6637bfa5 not found: ID does not exist" Oct 01 09:56:12 crc kubenswrapper[4787]: I1001 09:56:12.433120 4787 scope.go:117] "RemoveContainer" containerID="f1bc5cf1281c1f641d7c326c41bd88fcd42be40d931224664f09b2e9338df0cb" Oct 01 09:56:12 crc kubenswrapper[4787]: E1001 09:56:12.433691 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1bc5cf1281c1f641d7c326c41bd88fcd42be40d931224664f09b2e9338df0cb\": container with ID starting with f1bc5cf1281c1f641d7c326c41bd88fcd42be40d931224664f09b2e9338df0cb not found: ID does not exist" containerID="f1bc5cf1281c1f641d7c326c41bd88fcd42be40d931224664f09b2e9338df0cb" Oct 01 09:56:12 crc kubenswrapper[4787]: I1001 09:56:12.433715 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1bc5cf1281c1f641d7c326c41bd88fcd42be40d931224664f09b2e9338df0cb"} err="failed to get container status \"f1bc5cf1281c1f641d7c326c41bd88fcd42be40d931224664f09b2e9338df0cb\": rpc error: code = NotFound desc = could not find container \"f1bc5cf1281c1f641d7c326c41bd88fcd42be40d931224664f09b2e9338df0cb\": container with ID starting with f1bc5cf1281c1f641d7c326c41bd88fcd42be40d931224664f09b2e9338df0cb not found: ID does not exist" Oct 01 09:56:12 crc kubenswrapper[4787]: I1001 09:56:12.443030 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-68b59d98cf-xk49b"] Oct 01 09:56:12 crc kubenswrapper[4787]: I1001 09:56:12.536841 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed8acab0-e099-4b65-9883-5ea45cc4599e" path="/var/lib/kubelet/pods/ed8acab0-e099-4b65-9883-5ea45cc4599e/volumes" Oct 01 09:56:16 crc kubenswrapper[4787]: I1001 09:56:16.867244 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-68b59d98cf-xk49b" podUID="ed8acab0-e099-4b65-9883-5ea45cc4599e" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.201:5353: i/o timeout" Oct 01 09:56:20 crc kubenswrapper[4787]: I1001 09:56:20.489227 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wk6cx"] Oct 01 09:56:20 crc kubenswrapper[4787]: E1001 09:56:20.490309 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed8acab0-e099-4b65-9883-5ea45cc4599e" containerName="dnsmasq-dns" Oct 01 09:56:20 crc kubenswrapper[4787]: I1001 09:56:20.490324 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed8acab0-e099-4b65-9883-5ea45cc4599e" containerName="dnsmasq-dns" Oct 01 09:56:20 crc kubenswrapper[4787]: E1001 09:56:20.490344 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed8acab0-e099-4b65-9883-5ea45cc4599e" containerName="init" Oct 01 09:56:20 crc kubenswrapper[4787]: I1001 09:56:20.490352 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed8acab0-e099-4b65-9883-5ea45cc4599e" containerName="init" Oct 01 09:56:20 crc kubenswrapper[4787]: I1001 09:56:20.490616 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed8acab0-e099-4b65-9883-5ea45cc4599e" containerName="dnsmasq-dns" Oct 01 09:56:20 crc kubenswrapper[4787]: I1001 09:56:20.491536 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wk6cx" Oct 01 09:56:20 crc kubenswrapper[4787]: I1001 09:56:20.494193 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 09:56:20 crc kubenswrapper[4787]: I1001 09:56:20.494225 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 09:56:20 crc kubenswrapper[4787]: I1001 09:56:20.494701 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 09:56:20 crc kubenswrapper[4787]: I1001 09:56:20.494765 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vh5v6" Oct 01 09:56:20 crc kubenswrapper[4787]: I1001 09:56:20.496820 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wk6cx"] Oct 01 09:56:20 crc kubenswrapper[4787]: I1001 09:56:20.597857 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8ba2dbe-8a68-4da2-ac35-664a50fb5061-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wk6cx\" (UID: \"d8ba2dbe-8a68-4da2-ac35-664a50fb5061\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wk6cx" Oct 01 09:56:20 crc kubenswrapper[4787]: I1001 09:56:20.597934 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cl8qj\" (UniqueName: \"kubernetes.io/projected/d8ba2dbe-8a68-4da2-ac35-664a50fb5061-kube-api-access-cl8qj\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wk6cx\" (UID: \"d8ba2dbe-8a68-4da2-ac35-664a50fb5061\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wk6cx" Oct 01 09:56:20 crc kubenswrapper[4787]: I1001 09:56:20.597953 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8ba2dbe-8a68-4da2-ac35-664a50fb5061-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wk6cx\" (UID: \"d8ba2dbe-8a68-4da2-ac35-664a50fb5061\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wk6cx" Oct 01 09:56:20 crc kubenswrapper[4787]: I1001 09:56:20.597983 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8ba2dbe-8a68-4da2-ac35-664a50fb5061-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wk6cx\" (UID: \"d8ba2dbe-8a68-4da2-ac35-664a50fb5061\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wk6cx" Oct 01 09:56:20 crc kubenswrapper[4787]: I1001 09:56:20.699756 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8ba2dbe-8a68-4da2-ac35-664a50fb5061-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wk6cx\" (UID: \"d8ba2dbe-8a68-4da2-ac35-664a50fb5061\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wk6cx" Oct 01 09:56:20 crc kubenswrapper[4787]: I1001 09:56:20.700016 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cl8qj\" (UniqueName: \"kubernetes.io/projected/d8ba2dbe-8a68-4da2-ac35-664a50fb5061-kube-api-access-cl8qj\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wk6cx\" (UID: \"d8ba2dbe-8a68-4da2-ac35-664a50fb5061\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wk6cx" Oct 01 09:56:20 crc kubenswrapper[4787]: I1001 09:56:20.700119 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8ba2dbe-8a68-4da2-ac35-664a50fb5061-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wk6cx\" (UID: \"d8ba2dbe-8a68-4da2-ac35-664a50fb5061\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wk6cx" Oct 01 09:56:20 crc kubenswrapper[4787]: I1001 09:56:20.700248 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8ba2dbe-8a68-4da2-ac35-664a50fb5061-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wk6cx\" (UID: \"d8ba2dbe-8a68-4da2-ac35-664a50fb5061\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wk6cx" Oct 01 09:56:20 crc kubenswrapper[4787]: I1001 09:56:20.705794 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8ba2dbe-8a68-4da2-ac35-664a50fb5061-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wk6cx\" (UID: \"d8ba2dbe-8a68-4da2-ac35-664a50fb5061\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wk6cx" Oct 01 09:56:20 crc kubenswrapper[4787]: I1001 09:56:20.705998 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8ba2dbe-8a68-4da2-ac35-664a50fb5061-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wk6cx\" (UID: \"d8ba2dbe-8a68-4da2-ac35-664a50fb5061\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wk6cx" Oct 01 09:56:20 crc kubenswrapper[4787]: I1001 09:56:20.707787 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8ba2dbe-8a68-4da2-ac35-664a50fb5061-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wk6cx\" (UID: \"d8ba2dbe-8a68-4da2-ac35-664a50fb5061\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wk6cx" Oct 01 09:56:20 crc kubenswrapper[4787]: I1001 09:56:20.723137 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cl8qj\" (UniqueName: \"kubernetes.io/projected/d8ba2dbe-8a68-4da2-ac35-664a50fb5061-kube-api-access-cl8qj\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wk6cx\" (UID: \"d8ba2dbe-8a68-4da2-ac35-664a50fb5061\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wk6cx" Oct 01 09:56:20 crc kubenswrapper[4787]: I1001 09:56:20.814270 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wk6cx" Oct 01 09:56:21 crc kubenswrapper[4787]: I1001 09:56:21.350218 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wk6cx"] Oct 01 09:56:21 crc kubenswrapper[4787]: I1001 09:56:21.358903 4787 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 09:56:21 crc kubenswrapper[4787]: I1001 09:56:21.455766 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wk6cx" event={"ID":"d8ba2dbe-8a68-4da2-ac35-664a50fb5061","Type":"ContainerStarted","Data":"3b75c94dccc3f856454874e4c9e70a2e088efa63d5bae9c12d94338495cf6df8"} Oct 01 09:56:30 crc kubenswrapper[4787]: I1001 09:56:30.563061 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wk6cx" event={"ID":"d8ba2dbe-8a68-4da2-ac35-664a50fb5061","Type":"ContainerStarted","Data":"24ec079883f5e3a74990556a153dbbab2d2cab02d8d19e6863b909a4e18e6840"} Oct 01 09:56:30 crc kubenswrapper[4787]: I1001 09:56:30.593693 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wk6cx" podStartSLOduration=2.131973685 podStartE2EDuration="10.5936718s" podCreationTimestamp="2025-10-01 09:56:20 +0000 UTC" firstStartedPulling="2025-10-01 09:56:21.358643868 +0000 UTC m=+1213.473788025" lastFinishedPulling="2025-10-01 09:56:29.820341963 +0000 UTC m=+1221.935486140" observedRunningTime="2025-10-01 09:56:30.584333102 +0000 UTC m=+1222.699477289" watchObservedRunningTime="2025-10-01 09:56:30.5936718 +0000 UTC m=+1222.708815967" Oct 01 09:56:35 crc kubenswrapper[4787]: I1001 09:56:35.611411 4787 generic.go:334] "Generic (PLEG): container finished" podID="eb5022f8-276b-4d6d-917b-be96a1d993bf" containerID="8043fd4f5d9a066ef9a0029929fd72592afd40dbb50a23407e5a155c8f2364aa" exitCode=0 Oct 01 09:56:35 crc kubenswrapper[4787]: I1001 09:56:35.611502 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"eb5022f8-276b-4d6d-917b-be96a1d993bf","Type":"ContainerDied","Data":"8043fd4f5d9a066ef9a0029929fd72592afd40dbb50a23407e5a155c8f2364aa"} Oct 01 09:56:36 crc kubenswrapper[4787]: I1001 09:56:36.624825 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"eb5022f8-276b-4d6d-917b-be96a1d993bf","Type":"ContainerStarted","Data":"2de863038ab6dbf07b0bbd3381e6c8a65a3d7d4d6e31f723d77d3141ba96abf3"} Oct 01 09:56:36 crc kubenswrapper[4787]: I1001 09:56:36.625391 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 01 09:56:36 crc kubenswrapper[4787]: I1001 09:56:36.626271 4787 generic.go:334] "Generic (PLEG): container finished" podID="d0353483-c978-4d9a-9c34-73ee379325ba" containerID="03af902c2fd8ec77a8dd910f7e1f4b6f157e978fef61cc57265b74c68df142b4" exitCode=0 Oct 01 09:56:36 crc kubenswrapper[4787]: I1001 09:56:36.626304 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d0353483-c978-4d9a-9c34-73ee379325ba","Type":"ContainerDied","Data":"03af902c2fd8ec77a8dd910f7e1f4b6f157e978fef61cc57265b74c68df142b4"} Oct 01 09:56:36 crc kubenswrapper[4787]: I1001 09:56:36.682194 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.682172074 podStartE2EDuration="37.682172074s" podCreationTimestamp="2025-10-01 09:55:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:56:36.663489549 +0000 UTC m=+1228.778633736" watchObservedRunningTime="2025-10-01 09:56:36.682172074 +0000 UTC m=+1228.797316241" Oct 01 09:56:37 crc kubenswrapper[4787]: I1001 09:56:37.645838 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d0353483-c978-4d9a-9c34-73ee379325ba","Type":"ContainerStarted","Data":"81ec61fe5e89d4017e2632e35ea73f8efd22cf7112498d9c2a76a1ea1cb34d80"} Oct 01 09:56:37 crc kubenswrapper[4787]: I1001 09:56:37.678689 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.678657184 podStartE2EDuration="38.678657184s" podCreationTimestamp="2025-10-01 09:55:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 09:56:37.671623175 +0000 UTC m=+1229.786767342" watchObservedRunningTime="2025-10-01 09:56:37.678657184 +0000 UTC m=+1229.793801351" Oct 01 09:56:39 crc kubenswrapper[4787]: I1001 09:56:39.645850 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:56:41 crc kubenswrapper[4787]: I1001 09:56:41.250564 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:56:41 crc kubenswrapper[4787]: I1001 09:56:41.250951 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:56:41 crc kubenswrapper[4787]: I1001 09:56:41.689469 4787 generic.go:334] "Generic (PLEG): container finished" podID="d8ba2dbe-8a68-4da2-ac35-664a50fb5061" containerID="24ec079883f5e3a74990556a153dbbab2d2cab02d8d19e6863b909a4e18e6840" exitCode=0 Oct 01 09:56:41 crc kubenswrapper[4787]: I1001 09:56:41.689552 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wk6cx" event={"ID":"d8ba2dbe-8a68-4da2-ac35-664a50fb5061","Type":"ContainerDied","Data":"24ec079883f5e3a74990556a153dbbab2d2cab02d8d19e6863b909a4e18e6840"} Oct 01 09:56:43 crc kubenswrapper[4787]: I1001 09:56:43.214833 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wk6cx" Oct 01 09:56:43 crc kubenswrapper[4787]: I1001 09:56:43.280524 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8ba2dbe-8a68-4da2-ac35-664a50fb5061-ssh-key\") pod \"d8ba2dbe-8a68-4da2-ac35-664a50fb5061\" (UID: \"d8ba2dbe-8a68-4da2-ac35-664a50fb5061\") " Oct 01 09:56:43 crc kubenswrapper[4787]: I1001 09:56:43.280710 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cl8qj\" (UniqueName: \"kubernetes.io/projected/d8ba2dbe-8a68-4da2-ac35-664a50fb5061-kube-api-access-cl8qj\") pod \"d8ba2dbe-8a68-4da2-ac35-664a50fb5061\" (UID: \"d8ba2dbe-8a68-4da2-ac35-664a50fb5061\") " Oct 01 09:56:43 crc kubenswrapper[4787]: I1001 09:56:43.280860 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8ba2dbe-8a68-4da2-ac35-664a50fb5061-repo-setup-combined-ca-bundle\") pod \"d8ba2dbe-8a68-4da2-ac35-664a50fb5061\" (UID: \"d8ba2dbe-8a68-4da2-ac35-664a50fb5061\") " Oct 01 09:56:43 crc kubenswrapper[4787]: I1001 09:56:43.280988 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8ba2dbe-8a68-4da2-ac35-664a50fb5061-inventory\") pod \"d8ba2dbe-8a68-4da2-ac35-664a50fb5061\" (UID: \"d8ba2dbe-8a68-4da2-ac35-664a50fb5061\") " Oct 01 09:56:43 crc kubenswrapper[4787]: I1001 09:56:43.287142 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8ba2dbe-8a68-4da2-ac35-664a50fb5061-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "d8ba2dbe-8a68-4da2-ac35-664a50fb5061" (UID: "d8ba2dbe-8a68-4da2-ac35-664a50fb5061"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:56:43 crc kubenswrapper[4787]: I1001 09:56:43.288408 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8ba2dbe-8a68-4da2-ac35-664a50fb5061-kube-api-access-cl8qj" (OuterVolumeSpecName: "kube-api-access-cl8qj") pod "d8ba2dbe-8a68-4da2-ac35-664a50fb5061" (UID: "d8ba2dbe-8a68-4da2-ac35-664a50fb5061"). InnerVolumeSpecName "kube-api-access-cl8qj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:56:43 crc kubenswrapper[4787]: I1001 09:56:43.313247 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8ba2dbe-8a68-4da2-ac35-664a50fb5061-inventory" (OuterVolumeSpecName: "inventory") pod "d8ba2dbe-8a68-4da2-ac35-664a50fb5061" (UID: "d8ba2dbe-8a68-4da2-ac35-664a50fb5061"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:56:43 crc kubenswrapper[4787]: I1001 09:56:43.316424 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8ba2dbe-8a68-4da2-ac35-664a50fb5061-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d8ba2dbe-8a68-4da2-ac35-664a50fb5061" (UID: "d8ba2dbe-8a68-4da2-ac35-664a50fb5061"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:56:43 crc kubenswrapper[4787]: I1001 09:56:43.383173 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cl8qj\" (UniqueName: \"kubernetes.io/projected/d8ba2dbe-8a68-4da2-ac35-664a50fb5061-kube-api-access-cl8qj\") on node \"crc\" DevicePath \"\"" Oct 01 09:56:43 crc kubenswrapper[4787]: I1001 09:56:43.383213 4787 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8ba2dbe-8a68-4da2-ac35-664a50fb5061-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:56:43 crc kubenswrapper[4787]: I1001 09:56:43.383226 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8ba2dbe-8a68-4da2-ac35-664a50fb5061-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 09:56:43 crc kubenswrapper[4787]: I1001 09:56:43.383239 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8ba2dbe-8a68-4da2-ac35-664a50fb5061-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:56:43 crc kubenswrapper[4787]: I1001 09:56:43.732898 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wk6cx" event={"ID":"d8ba2dbe-8a68-4da2-ac35-664a50fb5061","Type":"ContainerDied","Data":"3b75c94dccc3f856454874e4c9e70a2e088efa63d5bae9c12d94338495cf6df8"} Oct 01 09:56:43 crc kubenswrapper[4787]: I1001 09:56:43.732952 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b75c94dccc3f856454874e4c9e70a2e088efa63d5bae9c12d94338495cf6df8" Oct 01 09:56:43 crc kubenswrapper[4787]: I1001 09:56:43.733019 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wk6cx" Oct 01 09:56:43 crc kubenswrapper[4787]: I1001 09:56:43.895672 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-j5ltq"] Oct 01 09:56:43 crc kubenswrapper[4787]: E1001 09:56:43.896910 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8ba2dbe-8a68-4da2-ac35-664a50fb5061" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 01 09:56:43 crc kubenswrapper[4787]: I1001 09:56:43.896941 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8ba2dbe-8a68-4da2-ac35-664a50fb5061" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 01 09:56:43 crc kubenswrapper[4787]: I1001 09:56:43.897223 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8ba2dbe-8a68-4da2-ac35-664a50fb5061" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 01 09:56:43 crc kubenswrapper[4787]: I1001 09:56:43.898127 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-j5ltq" Oct 01 09:56:43 crc kubenswrapper[4787]: I1001 09:56:43.900394 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 09:56:43 crc kubenswrapper[4787]: I1001 09:56:43.900632 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vh5v6" Oct 01 09:56:43 crc kubenswrapper[4787]: I1001 09:56:43.900884 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 09:56:43 crc kubenswrapper[4787]: I1001 09:56:43.901992 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 09:56:43 crc kubenswrapper[4787]: I1001 09:56:43.913418 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-j5ltq"] Oct 01 09:56:43 crc kubenswrapper[4787]: I1001 09:56:43.998899 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vn79w\" (UniqueName: \"kubernetes.io/projected/8d46852a-ab64-4e0c-b29a-e9bde8305a94-kube-api-access-vn79w\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-j5ltq\" (UID: \"8d46852a-ab64-4e0c-b29a-e9bde8305a94\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-j5ltq" Oct 01 09:56:43 crc kubenswrapper[4787]: I1001 09:56:43.999417 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8d46852a-ab64-4e0c-b29a-e9bde8305a94-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-j5ltq\" (UID: \"8d46852a-ab64-4e0c-b29a-e9bde8305a94\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-j5ltq" Oct 01 09:56:43 crc kubenswrapper[4787]: I1001 09:56:43.999553 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8d46852a-ab64-4e0c-b29a-e9bde8305a94-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-j5ltq\" (UID: \"8d46852a-ab64-4e0c-b29a-e9bde8305a94\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-j5ltq" Oct 01 09:56:44 crc kubenswrapper[4787]: I1001 09:56:44.101631 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8d46852a-ab64-4e0c-b29a-e9bde8305a94-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-j5ltq\" (UID: \"8d46852a-ab64-4e0c-b29a-e9bde8305a94\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-j5ltq" Oct 01 09:56:44 crc kubenswrapper[4787]: I1001 09:56:44.101724 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8d46852a-ab64-4e0c-b29a-e9bde8305a94-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-j5ltq\" (UID: \"8d46852a-ab64-4e0c-b29a-e9bde8305a94\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-j5ltq" Oct 01 09:56:44 crc kubenswrapper[4787]: I1001 09:56:44.101800 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vn79w\" (UniqueName: \"kubernetes.io/projected/8d46852a-ab64-4e0c-b29a-e9bde8305a94-kube-api-access-vn79w\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-j5ltq\" (UID: \"8d46852a-ab64-4e0c-b29a-e9bde8305a94\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-j5ltq" Oct 01 09:56:44 crc kubenswrapper[4787]: I1001 09:56:44.119524 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8d46852a-ab64-4e0c-b29a-e9bde8305a94-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-j5ltq\" (UID: \"8d46852a-ab64-4e0c-b29a-e9bde8305a94\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-j5ltq" Oct 01 09:56:44 crc kubenswrapper[4787]: I1001 09:56:44.122373 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8d46852a-ab64-4e0c-b29a-e9bde8305a94-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-j5ltq\" (UID: \"8d46852a-ab64-4e0c-b29a-e9bde8305a94\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-j5ltq" Oct 01 09:56:44 crc kubenswrapper[4787]: I1001 09:56:44.122559 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vn79w\" (UniqueName: \"kubernetes.io/projected/8d46852a-ab64-4e0c-b29a-e9bde8305a94-kube-api-access-vn79w\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-j5ltq\" (UID: \"8d46852a-ab64-4e0c-b29a-e9bde8305a94\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-j5ltq" Oct 01 09:56:44 crc kubenswrapper[4787]: I1001 09:56:44.218085 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-j5ltq" Oct 01 09:56:44 crc kubenswrapper[4787]: I1001 09:56:44.835807 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-j5ltq"] Oct 01 09:56:45 crc kubenswrapper[4787]: I1001 09:56:45.756917 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-j5ltq" event={"ID":"8d46852a-ab64-4e0c-b29a-e9bde8305a94","Type":"ContainerStarted","Data":"878e18c16016d8aa4c085c33d43e3c929b8e6a1674d9e64892777eb6bd39b5b5"} Oct 01 09:56:45 crc kubenswrapper[4787]: I1001 09:56:45.757285 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-j5ltq" event={"ID":"8d46852a-ab64-4e0c-b29a-e9bde8305a94","Type":"ContainerStarted","Data":"084eb5ae250b7c7e190674317b258f490d9905a3b3dae1a646925f08854b8a6a"} Oct 01 09:56:45 crc kubenswrapper[4787]: I1001 09:56:45.801811 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-j5ltq" podStartSLOduration=2.401432223 podStartE2EDuration="2.801793756s" podCreationTimestamp="2025-10-01 09:56:43 +0000 UTC" firstStartedPulling="2025-10-01 09:56:44.854244251 +0000 UTC m=+1236.969388408" lastFinishedPulling="2025-10-01 09:56:45.254605784 +0000 UTC m=+1237.369749941" observedRunningTime="2025-10-01 09:56:45.790827591 +0000 UTC m=+1237.905971748" watchObservedRunningTime="2025-10-01 09:56:45.801793756 +0000 UTC m=+1237.916937913" Oct 01 09:56:48 crc kubenswrapper[4787]: I1001 09:56:48.800469 4787 generic.go:334] "Generic (PLEG): container finished" podID="8d46852a-ab64-4e0c-b29a-e9bde8305a94" containerID="878e18c16016d8aa4c085c33d43e3c929b8e6a1674d9e64892777eb6bd39b5b5" exitCode=0 Oct 01 09:56:48 crc kubenswrapper[4787]: I1001 09:56:48.800559 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-j5ltq" event={"ID":"8d46852a-ab64-4e0c-b29a-e9bde8305a94","Type":"ContainerDied","Data":"878e18c16016d8aa4c085c33d43e3c929b8e6a1674d9e64892777eb6bd39b5b5"} Oct 01 09:56:49 crc kubenswrapper[4787]: I1001 09:56:49.587419 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 01 09:56:49 crc kubenswrapper[4787]: I1001 09:56:49.651476 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 01 09:56:50 crc kubenswrapper[4787]: I1001 09:56:50.397780 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-j5ltq" Oct 01 09:56:50 crc kubenswrapper[4787]: I1001 09:56:50.448172 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8d46852a-ab64-4e0c-b29a-e9bde8305a94-inventory\") pod \"8d46852a-ab64-4e0c-b29a-e9bde8305a94\" (UID: \"8d46852a-ab64-4e0c-b29a-e9bde8305a94\") " Oct 01 09:56:50 crc kubenswrapper[4787]: I1001 09:56:50.448277 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vn79w\" (UniqueName: \"kubernetes.io/projected/8d46852a-ab64-4e0c-b29a-e9bde8305a94-kube-api-access-vn79w\") pod \"8d46852a-ab64-4e0c-b29a-e9bde8305a94\" (UID: \"8d46852a-ab64-4e0c-b29a-e9bde8305a94\") " Oct 01 09:56:50 crc kubenswrapper[4787]: I1001 09:56:50.448516 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8d46852a-ab64-4e0c-b29a-e9bde8305a94-ssh-key\") pod \"8d46852a-ab64-4e0c-b29a-e9bde8305a94\" (UID: \"8d46852a-ab64-4e0c-b29a-e9bde8305a94\") " Oct 01 09:56:50 crc kubenswrapper[4787]: I1001 09:56:50.460289 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d46852a-ab64-4e0c-b29a-e9bde8305a94-kube-api-access-vn79w" (OuterVolumeSpecName: "kube-api-access-vn79w") pod "8d46852a-ab64-4e0c-b29a-e9bde8305a94" (UID: "8d46852a-ab64-4e0c-b29a-e9bde8305a94"). InnerVolumeSpecName "kube-api-access-vn79w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:56:50 crc kubenswrapper[4787]: I1001 09:56:50.487801 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d46852a-ab64-4e0c-b29a-e9bde8305a94-inventory" (OuterVolumeSpecName: "inventory") pod "8d46852a-ab64-4e0c-b29a-e9bde8305a94" (UID: "8d46852a-ab64-4e0c-b29a-e9bde8305a94"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:56:50 crc kubenswrapper[4787]: I1001 09:56:50.487803 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d46852a-ab64-4e0c-b29a-e9bde8305a94-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8d46852a-ab64-4e0c-b29a-e9bde8305a94" (UID: "8d46852a-ab64-4e0c-b29a-e9bde8305a94"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:56:50 crc kubenswrapper[4787]: I1001 09:56:50.551750 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8d46852a-ab64-4e0c-b29a-e9bde8305a94-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 09:56:50 crc kubenswrapper[4787]: I1001 09:56:50.551856 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vn79w\" (UniqueName: \"kubernetes.io/projected/8d46852a-ab64-4e0c-b29a-e9bde8305a94-kube-api-access-vn79w\") on node \"crc\" DevicePath \"\"" Oct 01 09:56:50 crc kubenswrapper[4787]: I1001 09:56:50.551953 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8d46852a-ab64-4e0c-b29a-e9bde8305a94-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:56:50 crc kubenswrapper[4787]: I1001 09:56:50.847563 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-j5ltq" event={"ID":"8d46852a-ab64-4e0c-b29a-e9bde8305a94","Type":"ContainerDied","Data":"084eb5ae250b7c7e190674317b258f490d9905a3b3dae1a646925f08854b8a6a"} Oct 01 09:56:50 crc kubenswrapper[4787]: I1001 09:56:50.848281 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="084eb5ae250b7c7e190674317b258f490d9905a3b3dae1a646925f08854b8a6a" Oct 01 09:56:50 crc kubenswrapper[4787]: I1001 09:56:50.848415 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-j5ltq" Oct 01 09:56:50 crc kubenswrapper[4787]: I1001 09:56:50.908966 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s4tf6"] Oct 01 09:56:50 crc kubenswrapper[4787]: E1001 09:56:50.909540 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d46852a-ab64-4e0c-b29a-e9bde8305a94" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 01 09:56:50 crc kubenswrapper[4787]: I1001 09:56:50.909557 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d46852a-ab64-4e0c-b29a-e9bde8305a94" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 01 09:56:50 crc kubenswrapper[4787]: I1001 09:56:50.909792 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d46852a-ab64-4e0c-b29a-e9bde8305a94" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 01 09:56:50 crc kubenswrapper[4787]: I1001 09:56:50.910646 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s4tf6" Oct 01 09:56:50 crc kubenswrapper[4787]: I1001 09:56:50.917219 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 09:56:50 crc kubenswrapper[4787]: I1001 09:56:50.917265 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 09:56:50 crc kubenswrapper[4787]: I1001 09:56:50.917387 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vh5v6" Oct 01 09:56:50 crc kubenswrapper[4787]: I1001 09:56:50.917613 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 09:56:50 crc kubenswrapper[4787]: I1001 09:56:50.929563 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s4tf6"] Oct 01 09:56:50 crc kubenswrapper[4787]: I1001 09:56:50.965321 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1018ece-ee61-4913-bd2f-64f996ef135c-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-s4tf6\" (UID: \"d1018ece-ee61-4913-bd2f-64f996ef135c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s4tf6" Oct 01 09:56:50 crc kubenswrapper[4787]: I1001 09:56:50.965376 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6h6dv\" (UniqueName: \"kubernetes.io/projected/d1018ece-ee61-4913-bd2f-64f996ef135c-kube-api-access-6h6dv\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-s4tf6\" (UID: \"d1018ece-ee61-4913-bd2f-64f996ef135c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s4tf6" Oct 01 09:56:50 crc kubenswrapper[4787]: I1001 09:56:50.965416 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d1018ece-ee61-4913-bd2f-64f996ef135c-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-s4tf6\" (UID: \"d1018ece-ee61-4913-bd2f-64f996ef135c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s4tf6" Oct 01 09:56:50 crc kubenswrapper[4787]: I1001 09:56:50.965536 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d1018ece-ee61-4913-bd2f-64f996ef135c-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-s4tf6\" (UID: \"d1018ece-ee61-4913-bd2f-64f996ef135c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s4tf6" Oct 01 09:56:51 crc kubenswrapper[4787]: I1001 09:56:51.067199 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d1018ece-ee61-4913-bd2f-64f996ef135c-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-s4tf6\" (UID: \"d1018ece-ee61-4913-bd2f-64f996ef135c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s4tf6" Oct 01 09:56:51 crc kubenswrapper[4787]: I1001 09:56:51.067362 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d1018ece-ee61-4913-bd2f-64f996ef135c-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-s4tf6\" (UID: \"d1018ece-ee61-4913-bd2f-64f996ef135c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s4tf6" Oct 01 09:56:51 crc kubenswrapper[4787]: I1001 09:56:51.067414 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1018ece-ee61-4913-bd2f-64f996ef135c-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-s4tf6\" (UID: \"d1018ece-ee61-4913-bd2f-64f996ef135c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s4tf6" Oct 01 09:56:51 crc kubenswrapper[4787]: I1001 09:56:51.067442 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6h6dv\" (UniqueName: \"kubernetes.io/projected/d1018ece-ee61-4913-bd2f-64f996ef135c-kube-api-access-6h6dv\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-s4tf6\" (UID: \"d1018ece-ee61-4913-bd2f-64f996ef135c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s4tf6" Oct 01 09:56:51 crc kubenswrapper[4787]: I1001 09:56:51.072177 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d1018ece-ee61-4913-bd2f-64f996ef135c-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-s4tf6\" (UID: \"d1018ece-ee61-4913-bd2f-64f996ef135c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s4tf6" Oct 01 09:56:51 crc kubenswrapper[4787]: I1001 09:56:51.072332 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d1018ece-ee61-4913-bd2f-64f996ef135c-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-s4tf6\" (UID: \"d1018ece-ee61-4913-bd2f-64f996ef135c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s4tf6" Oct 01 09:56:51 crc kubenswrapper[4787]: I1001 09:56:51.073948 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1018ece-ee61-4913-bd2f-64f996ef135c-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-s4tf6\" (UID: \"d1018ece-ee61-4913-bd2f-64f996ef135c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s4tf6" Oct 01 09:56:51 crc kubenswrapper[4787]: I1001 09:56:51.084235 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6h6dv\" (UniqueName: \"kubernetes.io/projected/d1018ece-ee61-4913-bd2f-64f996ef135c-kube-api-access-6h6dv\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-s4tf6\" (UID: \"d1018ece-ee61-4913-bd2f-64f996ef135c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s4tf6" Oct 01 09:56:51 crc kubenswrapper[4787]: I1001 09:56:51.230923 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s4tf6" Oct 01 09:56:51 crc kubenswrapper[4787]: W1001 09:56:51.810325 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd1018ece_ee61_4913_bd2f_64f996ef135c.slice/crio-1604dcca1e397a31c19754c01dd5f17702126d7b51b04c38aa226d4e07e39103 WatchSource:0}: Error finding container 1604dcca1e397a31c19754c01dd5f17702126d7b51b04c38aa226d4e07e39103: Status 404 returned error can't find the container with id 1604dcca1e397a31c19754c01dd5f17702126d7b51b04c38aa226d4e07e39103 Oct 01 09:56:51 crc kubenswrapper[4787]: I1001 09:56:51.811913 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s4tf6"] Oct 01 09:56:51 crc kubenswrapper[4787]: I1001 09:56:51.861780 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s4tf6" event={"ID":"d1018ece-ee61-4913-bd2f-64f996ef135c","Type":"ContainerStarted","Data":"1604dcca1e397a31c19754c01dd5f17702126d7b51b04c38aa226d4e07e39103"} Oct 01 09:56:52 crc kubenswrapper[4787]: I1001 09:56:52.874598 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s4tf6" event={"ID":"d1018ece-ee61-4913-bd2f-64f996ef135c","Type":"ContainerStarted","Data":"b8edd8c7dc76be74e1f30784b4e406d40f803fd07cbb20f81682e17580421ed7"} Oct 01 09:56:52 crc kubenswrapper[4787]: I1001 09:56:52.905624 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s4tf6" podStartSLOduration=2.5244446209999998 podStartE2EDuration="2.905598692s" podCreationTimestamp="2025-10-01 09:56:50 +0000 UTC" firstStartedPulling="2025-10-01 09:56:51.815575877 +0000 UTC m=+1243.930720044" lastFinishedPulling="2025-10-01 09:56:52.196729958 +0000 UTC m=+1244.311874115" observedRunningTime="2025-10-01 09:56:52.892491683 +0000 UTC m=+1245.007635880" watchObservedRunningTime="2025-10-01 09:56:52.905598692 +0000 UTC m=+1245.020742859" Oct 01 09:57:11 crc kubenswrapper[4787]: I1001 09:57:11.250999 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:57:11 crc kubenswrapper[4787]: I1001 09:57:11.254476 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:57:29 crc kubenswrapper[4787]: I1001 09:57:29.817930 4787 scope.go:117] "RemoveContainer" containerID="e88e716b091e96dfbeec718541083c5b8f296f8f98fea8a472a2917ee8c45f46" Oct 01 09:57:29 crc kubenswrapper[4787]: I1001 09:57:29.845447 4787 scope.go:117] "RemoveContainer" containerID="6020b30fd47f95aa07a6b93dfa13052657b3120e459f6fe4674a14f5b6486a9c" Oct 01 09:57:29 crc kubenswrapper[4787]: I1001 09:57:29.880353 4787 scope.go:117] "RemoveContainer" containerID="2c2684e2732e0784cacbce917a9aa23a444a150b54d2e11ce18ac1efd3f00bd0" Oct 01 09:57:29 crc kubenswrapper[4787]: I1001 09:57:29.900868 4787 scope.go:117] "RemoveContainer" containerID="6f2178ce4b22d3637acac7d52e9173e5f851ab77726fba536ab7ec09e3872032" Oct 01 09:57:41 crc kubenswrapper[4787]: I1001 09:57:41.251155 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:57:41 crc kubenswrapper[4787]: I1001 09:57:41.252531 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:57:41 crc kubenswrapper[4787]: I1001 09:57:41.252656 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" Oct 01 09:57:41 crc kubenswrapper[4787]: I1001 09:57:41.254491 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a85043035562ba4be7fcc084d6a9297752334c5b023237849bd5d1037f591d75"} pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 09:57:41 crc kubenswrapper[4787]: I1001 09:57:41.254839 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" containerID="cri-o://a85043035562ba4be7fcc084d6a9297752334c5b023237849bd5d1037f591d75" gracePeriod=600 Oct 01 09:57:41 crc kubenswrapper[4787]: I1001 09:57:41.463619 4787 generic.go:334] "Generic (PLEG): container finished" podID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerID="a85043035562ba4be7fcc084d6a9297752334c5b023237849bd5d1037f591d75" exitCode=0 Oct 01 09:57:41 crc kubenswrapper[4787]: I1001 09:57:41.463704 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" event={"ID":"fa51b883-5f3e-4141-9d4c-37704aac7718","Type":"ContainerDied","Data":"a85043035562ba4be7fcc084d6a9297752334c5b023237849bd5d1037f591d75"} Oct 01 09:57:41 crc kubenswrapper[4787]: I1001 09:57:41.463824 4787 scope.go:117] "RemoveContainer" containerID="86d3a42fa94bb1529366685be8d4c187b77d63f0ed73a16bbe3812ceaa99b7ae" Oct 01 09:57:42 crc kubenswrapper[4787]: I1001 09:57:42.483470 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" event={"ID":"fa51b883-5f3e-4141-9d4c-37704aac7718","Type":"ContainerStarted","Data":"e01333312cf21546b537b606d139d2d34e67fa8885c08f9ceb1f4c790ae4fd3a"} Oct 01 09:58:30 crc kubenswrapper[4787]: I1001 09:58:30.007801 4787 scope.go:117] "RemoveContainer" containerID="dc3a7667b021a77fbd7004c14e7eea8bcaffddf0b1788a0efc6e5c22d34ef255" Oct 01 09:58:30 crc kubenswrapper[4787]: I1001 09:58:30.051909 4787 scope.go:117] "RemoveContainer" containerID="02b4037afb172a39d39693d2e624c65b1326bc9b665ead63181fab2ff8180c5c" Oct 01 09:58:30 crc kubenswrapper[4787]: I1001 09:58:30.084390 4787 scope.go:117] "RemoveContainer" containerID="48a9a1fbf4d60825c9724f9ea3da0e4ff19da26f8e5bac0bf0f259bd88d689bf" Oct 01 09:58:30 crc kubenswrapper[4787]: I1001 09:58:30.155792 4787 scope.go:117] "RemoveContainer" containerID="df38295b98c61b89830a5e9c42a88dd5e5dbafc3ddb80edc86f22e38179d976b" Oct 01 09:59:11 crc kubenswrapper[4787]: I1001 09:59:11.485002 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9wlnv"] Oct 01 09:59:11 crc kubenswrapper[4787]: I1001 09:59:11.487447 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9wlnv" Oct 01 09:59:11 crc kubenswrapper[4787]: I1001 09:59:11.510655 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9wlnv"] Oct 01 09:59:11 crc kubenswrapper[4787]: I1001 09:59:11.574444 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1af17f67-b247-48f9-a000-83d626489e29-utilities\") pod \"community-operators-9wlnv\" (UID: \"1af17f67-b247-48f9-a000-83d626489e29\") " pod="openshift-marketplace/community-operators-9wlnv" Oct 01 09:59:11 crc kubenswrapper[4787]: I1001 09:59:11.574524 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqlt7\" (UniqueName: \"kubernetes.io/projected/1af17f67-b247-48f9-a000-83d626489e29-kube-api-access-zqlt7\") pod \"community-operators-9wlnv\" (UID: \"1af17f67-b247-48f9-a000-83d626489e29\") " pod="openshift-marketplace/community-operators-9wlnv" Oct 01 09:59:11 crc kubenswrapper[4787]: I1001 09:59:11.574708 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1af17f67-b247-48f9-a000-83d626489e29-catalog-content\") pod \"community-operators-9wlnv\" (UID: \"1af17f67-b247-48f9-a000-83d626489e29\") " pod="openshift-marketplace/community-operators-9wlnv" Oct 01 09:59:11 crc kubenswrapper[4787]: I1001 09:59:11.676869 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1af17f67-b247-48f9-a000-83d626489e29-catalog-content\") pod \"community-operators-9wlnv\" (UID: \"1af17f67-b247-48f9-a000-83d626489e29\") " pod="openshift-marketplace/community-operators-9wlnv" Oct 01 09:59:11 crc kubenswrapper[4787]: I1001 09:59:11.676942 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1af17f67-b247-48f9-a000-83d626489e29-utilities\") pod \"community-operators-9wlnv\" (UID: \"1af17f67-b247-48f9-a000-83d626489e29\") " pod="openshift-marketplace/community-operators-9wlnv" Oct 01 09:59:11 crc kubenswrapper[4787]: I1001 09:59:11.676977 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqlt7\" (UniqueName: \"kubernetes.io/projected/1af17f67-b247-48f9-a000-83d626489e29-kube-api-access-zqlt7\") pod \"community-operators-9wlnv\" (UID: \"1af17f67-b247-48f9-a000-83d626489e29\") " pod="openshift-marketplace/community-operators-9wlnv" Oct 01 09:59:11 crc kubenswrapper[4787]: I1001 09:59:11.677574 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1af17f67-b247-48f9-a000-83d626489e29-catalog-content\") pod \"community-operators-9wlnv\" (UID: \"1af17f67-b247-48f9-a000-83d626489e29\") " pod="openshift-marketplace/community-operators-9wlnv" Oct 01 09:59:11 crc kubenswrapper[4787]: I1001 09:59:11.677619 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1af17f67-b247-48f9-a000-83d626489e29-utilities\") pod \"community-operators-9wlnv\" (UID: \"1af17f67-b247-48f9-a000-83d626489e29\") " pod="openshift-marketplace/community-operators-9wlnv" Oct 01 09:59:11 crc kubenswrapper[4787]: I1001 09:59:11.703403 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqlt7\" (UniqueName: \"kubernetes.io/projected/1af17f67-b247-48f9-a000-83d626489e29-kube-api-access-zqlt7\") pod \"community-operators-9wlnv\" (UID: \"1af17f67-b247-48f9-a000-83d626489e29\") " pod="openshift-marketplace/community-operators-9wlnv" Oct 01 09:59:11 crc kubenswrapper[4787]: I1001 09:59:11.807526 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9wlnv" Oct 01 09:59:12 crc kubenswrapper[4787]: I1001 09:59:12.311640 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9wlnv"] Oct 01 09:59:12 crc kubenswrapper[4787]: I1001 09:59:12.592502 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wlnv" event={"ID":"1af17f67-b247-48f9-a000-83d626489e29","Type":"ContainerStarted","Data":"a2517e55a9f70c72bf653de0a2484ee9052fb67c4e91200adc4630dd1359f3ae"} Oct 01 09:59:13 crc kubenswrapper[4787]: I1001 09:59:13.604665 4787 generic.go:334] "Generic (PLEG): container finished" podID="1af17f67-b247-48f9-a000-83d626489e29" containerID="5086f2ed6a5667b1e238ea8c1571df544cb6a8d112e57029fb42054d27c54146" exitCode=0 Oct 01 09:59:13 crc kubenswrapper[4787]: I1001 09:59:13.604735 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wlnv" event={"ID":"1af17f67-b247-48f9-a000-83d626489e29","Type":"ContainerDied","Data":"5086f2ed6a5667b1e238ea8c1571df544cb6a8d112e57029fb42054d27c54146"} Oct 01 09:59:14 crc kubenswrapper[4787]: I1001 09:59:14.270777 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zwsdr"] Oct 01 09:59:14 crc kubenswrapper[4787]: I1001 09:59:14.273995 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zwsdr" Oct 01 09:59:14 crc kubenswrapper[4787]: I1001 09:59:14.281765 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zwsdr"] Oct 01 09:59:14 crc kubenswrapper[4787]: I1001 09:59:14.435044 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77mdj\" (UniqueName: \"kubernetes.io/projected/6d846f41-b98b-41fd-b027-fd764c01c021-kube-api-access-77mdj\") pod \"certified-operators-zwsdr\" (UID: \"6d846f41-b98b-41fd-b027-fd764c01c021\") " pod="openshift-marketplace/certified-operators-zwsdr" Oct 01 09:59:14 crc kubenswrapper[4787]: I1001 09:59:14.435171 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d846f41-b98b-41fd-b027-fd764c01c021-utilities\") pod \"certified-operators-zwsdr\" (UID: \"6d846f41-b98b-41fd-b027-fd764c01c021\") " pod="openshift-marketplace/certified-operators-zwsdr" Oct 01 09:59:14 crc kubenswrapper[4787]: I1001 09:59:14.435253 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d846f41-b98b-41fd-b027-fd764c01c021-catalog-content\") pod \"certified-operators-zwsdr\" (UID: \"6d846f41-b98b-41fd-b027-fd764c01c021\") " pod="openshift-marketplace/certified-operators-zwsdr" Oct 01 09:59:14 crc kubenswrapper[4787]: I1001 09:59:14.564961 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77mdj\" (UniqueName: \"kubernetes.io/projected/6d846f41-b98b-41fd-b027-fd764c01c021-kube-api-access-77mdj\") pod \"certified-operators-zwsdr\" (UID: \"6d846f41-b98b-41fd-b027-fd764c01c021\") " pod="openshift-marketplace/certified-operators-zwsdr" Oct 01 09:59:14 crc kubenswrapper[4787]: I1001 09:59:14.565376 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d846f41-b98b-41fd-b027-fd764c01c021-utilities\") pod \"certified-operators-zwsdr\" (UID: \"6d846f41-b98b-41fd-b027-fd764c01c021\") " pod="openshift-marketplace/certified-operators-zwsdr" Oct 01 09:59:14 crc kubenswrapper[4787]: I1001 09:59:14.565585 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d846f41-b98b-41fd-b027-fd764c01c021-catalog-content\") pod \"certified-operators-zwsdr\" (UID: \"6d846f41-b98b-41fd-b027-fd764c01c021\") " pod="openshift-marketplace/certified-operators-zwsdr" Oct 01 09:59:14 crc kubenswrapper[4787]: I1001 09:59:14.566257 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d846f41-b98b-41fd-b027-fd764c01c021-catalog-content\") pod \"certified-operators-zwsdr\" (UID: \"6d846f41-b98b-41fd-b027-fd764c01c021\") " pod="openshift-marketplace/certified-operators-zwsdr" Oct 01 09:59:14 crc kubenswrapper[4787]: I1001 09:59:14.566674 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d846f41-b98b-41fd-b027-fd764c01c021-utilities\") pod \"certified-operators-zwsdr\" (UID: \"6d846f41-b98b-41fd-b027-fd764c01c021\") " pod="openshift-marketplace/certified-operators-zwsdr" Oct 01 09:59:14 crc kubenswrapper[4787]: I1001 09:59:14.593834 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77mdj\" (UniqueName: \"kubernetes.io/projected/6d846f41-b98b-41fd-b027-fd764c01c021-kube-api-access-77mdj\") pod \"certified-operators-zwsdr\" (UID: \"6d846f41-b98b-41fd-b027-fd764c01c021\") " pod="openshift-marketplace/certified-operators-zwsdr" Oct 01 09:59:14 crc kubenswrapper[4787]: I1001 09:59:14.595946 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zwsdr" Oct 01 09:59:14 crc kubenswrapper[4787]: I1001 09:59:14.626654 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wlnv" event={"ID":"1af17f67-b247-48f9-a000-83d626489e29","Type":"ContainerStarted","Data":"dabc5f80071fc67c284fc0916993e81c84c38cec5872e196f0b7ca93648c0de6"} Oct 01 09:59:15 crc kubenswrapper[4787]: I1001 09:59:15.131561 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zwsdr"] Oct 01 09:59:15 crc kubenswrapper[4787]: W1001 09:59:15.133231 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6d846f41_b98b_41fd_b027_fd764c01c021.slice/crio-ee439afd56d3c0674acdb73c29821a6d08db687228b3b76a8545615239a69713 WatchSource:0}: Error finding container ee439afd56d3c0674acdb73c29821a6d08db687228b3b76a8545615239a69713: Status 404 returned error can't find the container with id ee439afd56d3c0674acdb73c29821a6d08db687228b3b76a8545615239a69713 Oct 01 09:59:15 crc kubenswrapper[4787]: I1001 09:59:15.635921 4787 generic.go:334] "Generic (PLEG): container finished" podID="1af17f67-b247-48f9-a000-83d626489e29" containerID="dabc5f80071fc67c284fc0916993e81c84c38cec5872e196f0b7ca93648c0de6" exitCode=0 Oct 01 09:59:15 crc kubenswrapper[4787]: I1001 09:59:15.636218 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wlnv" event={"ID":"1af17f67-b247-48f9-a000-83d626489e29","Type":"ContainerDied","Data":"dabc5f80071fc67c284fc0916993e81c84c38cec5872e196f0b7ca93648c0de6"} Oct 01 09:59:15 crc kubenswrapper[4787]: I1001 09:59:15.638802 4787 generic.go:334] "Generic (PLEG): container finished" podID="6d846f41-b98b-41fd-b027-fd764c01c021" containerID="810d10621dfc5cf0561dc2847098f182f30774f720d46ed5ae306e1f86da5b51" exitCode=0 Oct 01 09:59:15 crc kubenswrapper[4787]: I1001 09:59:15.638849 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zwsdr" event={"ID":"6d846f41-b98b-41fd-b027-fd764c01c021","Type":"ContainerDied","Data":"810d10621dfc5cf0561dc2847098f182f30774f720d46ed5ae306e1f86da5b51"} Oct 01 09:59:15 crc kubenswrapper[4787]: I1001 09:59:15.638876 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zwsdr" event={"ID":"6d846f41-b98b-41fd-b027-fd764c01c021","Type":"ContainerStarted","Data":"ee439afd56d3c0674acdb73c29821a6d08db687228b3b76a8545615239a69713"} Oct 01 09:59:16 crc kubenswrapper[4787]: I1001 09:59:16.655739 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wlnv" event={"ID":"1af17f67-b247-48f9-a000-83d626489e29","Type":"ContainerStarted","Data":"0fcccd9be09180c3d7e72d5abba1d21a44f7d1bf7dae6990b75b27a2483d80dc"} Oct 01 09:59:16 crc kubenswrapper[4787]: I1001 09:59:16.659487 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zwsdr" event={"ID":"6d846f41-b98b-41fd-b027-fd764c01c021","Type":"ContainerStarted","Data":"0ad5444728a5182df188d71b14576784ac47c71ee00c4a769c0b55c9d9409c56"} Oct 01 09:59:16 crc kubenswrapper[4787]: I1001 09:59:16.682365 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9wlnv" podStartSLOduration=3.200795339 podStartE2EDuration="5.68234518s" podCreationTimestamp="2025-10-01 09:59:11 +0000 UTC" firstStartedPulling="2025-10-01 09:59:13.606977586 +0000 UTC m=+1385.722121763" lastFinishedPulling="2025-10-01 09:59:16.088527447 +0000 UTC m=+1388.203671604" observedRunningTime="2025-10-01 09:59:16.674038168 +0000 UTC m=+1388.789182345" watchObservedRunningTime="2025-10-01 09:59:16.68234518 +0000 UTC m=+1388.797489337" Oct 01 09:59:17 crc kubenswrapper[4787]: I1001 09:59:17.673761 4787 generic.go:334] "Generic (PLEG): container finished" podID="6d846f41-b98b-41fd-b027-fd764c01c021" containerID="0ad5444728a5182df188d71b14576784ac47c71ee00c4a769c0b55c9d9409c56" exitCode=0 Oct 01 09:59:17 crc kubenswrapper[4787]: I1001 09:59:17.673862 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zwsdr" event={"ID":"6d846f41-b98b-41fd-b027-fd764c01c021","Type":"ContainerDied","Data":"0ad5444728a5182df188d71b14576784ac47c71ee00c4a769c0b55c9d9409c56"} Oct 01 09:59:21 crc kubenswrapper[4787]: I1001 09:59:21.706197 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zwsdr" event={"ID":"6d846f41-b98b-41fd-b027-fd764c01c021","Type":"ContainerStarted","Data":"a217e19f0304c1f5b923a84decadd1d7d323ad112afd080d16c4e377bed351e5"} Oct 01 09:59:21 crc kubenswrapper[4787]: I1001 09:59:21.727321 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zwsdr" podStartSLOduration=2.528481976 podStartE2EDuration="7.727299709s" podCreationTimestamp="2025-10-01 09:59:14 +0000 UTC" firstStartedPulling="2025-10-01 09:59:15.640583615 +0000 UTC m=+1387.755727772" lastFinishedPulling="2025-10-01 09:59:20.839401288 +0000 UTC m=+1392.954545505" observedRunningTime="2025-10-01 09:59:21.72367143 +0000 UTC m=+1393.838815607" watchObservedRunningTime="2025-10-01 09:59:21.727299709 +0000 UTC m=+1393.842443866" Oct 01 09:59:21 crc kubenswrapper[4787]: I1001 09:59:21.807957 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9wlnv" Oct 01 09:59:21 crc kubenswrapper[4787]: I1001 09:59:21.808807 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9wlnv" Oct 01 09:59:21 crc kubenswrapper[4787]: I1001 09:59:21.862845 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9wlnv" Oct 01 09:59:22 crc kubenswrapper[4787]: I1001 09:59:22.768430 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9wlnv" Oct 01 09:59:23 crc kubenswrapper[4787]: I1001 09:59:23.068306 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9wlnv"] Oct 01 09:59:24 crc kubenswrapper[4787]: I1001 09:59:24.597061 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zwsdr" Oct 01 09:59:24 crc kubenswrapper[4787]: I1001 09:59:24.597355 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zwsdr" Oct 01 09:59:24 crc kubenswrapper[4787]: I1001 09:59:24.664999 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zwsdr" Oct 01 09:59:24 crc kubenswrapper[4787]: I1001 09:59:24.733202 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9wlnv" podUID="1af17f67-b247-48f9-a000-83d626489e29" containerName="registry-server" containerID="cri-o://0fcccd9be09180c3d7e72d5abba1d21a44f7d1bf7dae6990b75b27a2483d80dc" gracePeriod=2 Oct 01 09:59:25 crc kubenswrapper[4787]: I1001 09:59:25.209332 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9wlnv" Oct 01 09:59:25 crc kubenswrapper[4787]: I1001 09:59:25.316356 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1af17f67-b247-48f9-a000-83d626489e29-catalog-content\") pod \"1af17f67-b247-48f9-a000-83d626489e29\" (UID: \"1af17f67-b247-48f9-a000-83d626489e29\") " Oct 01 09:59:25 crc kubenswrapper[4787]: I1001 09:59:25.316594 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1af17f67-b247-48f9-a000-83d626489e29-utilities\") pod \"1af17f67-b247-48f9-a000-83d626489e29\" (UID: \"1af17f67-b247-48f9-a000-83d626489e29\") " Oct 01 09:59:25 crc kubenswrapper[4787]: I1001 09:59:25.316719 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zqlt7\" (UniqueName: \"kubernetes.io/projected/1af17f67-b247-48f9-a000-83d626489e29-kube-api-access-zqlt7\") pod \"1af17f67-b247-48f9-a000-83d626489e29\" (UID: \"1af17f67-b247-48f9-a000-83d626489e29\") " Oct 01 09:59:25 crc kubenswrapper[4787]: I1001 09:59:25.318776 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1af17f67-b247-48f9-a000-83d626489e29-utilities" (OuterVolumeSpecName: "utilities") pod "1af17f67-b247-48f9-a000-83d626489e29" (UID: "1af17f67-b247-48f9-a000-83d626489e29"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:59:25 crc kubenswrapper[4787]: I1001 09:59:25.326098 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1af17f67-b247-48f9-a000-83d626489e29-kube-api-access-zqlt7" (OuterVolumeSpecName: "kube-api-access-zqlt7") pod "1af17f67-b247-48f9-a000-83d626489e29" (UID: "1af17f67-b247-48f9-a000-83d626489e29"). InnerVolumeSpecName "kube-api-access-zqlt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:59:25 crc kubenswrapper[4787]: I1001 09:59:25.374470 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1af17f67-b247-48f9-a000-83d626489e29-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1af17f67-b247-48f9-a000-83d626489e29" (UID: "1af17f67-b247-48f9-a000-83d626489e29"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:59:25 crc kubenswrapper[4787]: I1001 09:59:25.419831 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1af17f67-b247-48f9-a000-83d626489e29-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:59:25 crc kubenswrapper[4787]: I1001 09:59:25.419871 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1af17f67-b247-48f9-a000-83d626489e29-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:59:25 crc kubenswrapper[4787]: I1001 09:59:25.419884 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zqlt7\" (UniqueName: \"kubernetes.io/projected/1af17f67-b247-48f9-a000-83d626489e29-kube-api-access-zqlt7\") on node \"crc\" DevicePath \"\"" Oct 01 09:59:25 crc kubenswrapper[4787]: I1001 09:59:25.744302 4787 generic.go:334] "Generic (PLEG): container finished" podID="1af17f67-b247-48f9-a000-83d626489e29" containerID="0fcccd9be09180c3d7e72d5abba1d21a44f7d1bf7dae6990b75b27a2483d80dc" exitCode=0 Oct 01 09:59:25 crc kubenswrapper[4787]: I1001 09:59:25.744350 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wlnv" event={"ID":"1af17f67-b247-48f9-a000-83d626489e29","Type":"ContainerDied","Data":"0fcccd9be09180c3d7e72d5abba1d21a44f7d1bf7dae6990b75b27a2483d80dc"} Oct 01 09:59:25 crc kubenswrapper[4787]: I1001 09:59:25.744383 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wlnv" event={"ID":"1af17f67-b247-48f9-a000-83d626489e29","Type":"ContainerDied","Data":"a2517e55a9f70c72bf653de0a2484ee9052fb67c4e91200adc4630dd1359f3ae"} Oct 01 09:59:25 crc kubenswrapper[4787]: I1001 09:59:25.744396 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9wlnv" Oct 01 09:59:25 crc kubenswrapper[4787]: I1001 09:59:25.744405 4787 scope.go:117] "RemoveContainer" containerID="0fcccd9be09180c3d7e72d5abba1d21a44f7d1bf7dae6990b75b27a2483d80dc" Oct 01 09:59:25 crc kubenswrapper[4787]: I1001 09:59:25.780122 4787 scope.go:117] "RemoveContainer" containerID="dabc5f80071fc67c284fc0916993e81c84c38cec5872e196f0b7ca93648c0de6" Oct 01 09:59:25 crc kubenswrapper[4787]: I1001 09:59:25.787524 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9wlnv"] Oct 01 09:59:25 crc kubenswrapper[4787]: I1001 09:59:25.811302 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9wlnv"] Oct 01 09:59:25 crc kubenswrapper[4787]: I1001 09:59:25.819284 4787 scope.go:117] "RemoveContainer" containerID="5086f2ed6a5667b1e238ea8c1571df544cb6a8d112e57029fb42054d27c54146" Oct 01 09:59:25 crc kubenswrapper[4787]: I1001 09:59:25.848278 4787 scope.go:117] "RemoveContainer" containerID="0fcccd9be09180c3d7e72d5abba1d21a44f7d1bf7dae6990b75b27a2483d80dc" Oct 01 09:59:25 crc kubenswrapper[4787]: E1001 09:59:25.848681 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fcccd9be09180c3d7e72d5abba1d21a44f7d1bf7dae6990b75b27a2483d80dc\": container with ID starting with 0fcccd9be09180c3d7e72d5abba1d21a44f7d1bf7dae6990b75b27a2483d80dc not found: ID does not exist" containerID="0fcccd9be09180c3d7e72d5abba1d21a44f7d1bf7dae6990b75b27a2483d80dc" Oct 01 09:59:25 crc kubenswrapper[4787]: I1001 09:59:25.848715 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fcccd9be09180c3d7e72d5abba1d21a44f7d1bf7dae6990b75b27a2483d80dc"} err="failed to get container status \"0fcccd9be09180c3d7e72d5abba1d21a44f7d1bf7dae6990b75b27a2483d80dc\": rpc error: code = NotFound desc = could not find container \"0fcccd9be09180c3d7e72d5abba1d21a44f7d1bf7dae6990b75b27a2483d80dc\": container with ID starting with 0fcccd9be09180c3d7e72d5abba1d21a44f7d1bf7dae6990b75b27a2483d80dc not found: ID does not exist" Oct 01 09:59:25 crc kubenswrapper[4787]: I1001 09:59:25.848743 4787 scope.go:117] "RemoveContainer" containerID="dabc5f80071fc67c284fc0916993e81c84c38cec5872e196f0b7ca93648c0de6" Oct 01 09:59:25 crc kubenswrapper[4787]: E1001 09:59:25.848983 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dabc5f80071fc67c284fc0916993e81c84c38cec5872e196f0b7ca93648c0de6\": container with ID starting with dabc5f80071fc67c284fc0916993e81c84c38cec5872e196f0b7ca93648c0de6 not found: ID does not exist" containerID="dabc5f80071fc67c284fc0916993e81c84c38cec5872e196f0b7ca93648c0de6" Oct 01 09:59:25 crc kubenswrapper[4787]: I1001 09:59:25.849005 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dabc5f80071fc67c284fc0916993e81c84c38cec5872e196f0b7ca93648c0de6"} err="failed to get container status \"dabc5f80071fc67c284fc0916993e81c84c38cec5872e196f0b7ca93648c0de6\": rpc error: code = NotFound desc = could not find container \"dabc5f80071fc67c284fc0916993e81c84c38cec5872e196f0b7ca93648c0de6\": container with ID starting with dabc5f80071fc67c284fc0916993e81c84c38cec5872e196f0b7ca93648c0de6 not found: ID does not exist" Oct 01 09:59:25 crc kubenswrapper[4787]: I1001 09:59:25.849022 4787 scope.go:117] "RemoveContainer" containerID="5086f2ed6a5667b1e238ea8c1571df544cb6a8d112e57029fb42054d27c54146" Oct 01 09:59:25 crc kubenswrapper[4787]: E1001 09:59:25.849405 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5086f2ed6a5667b1e238ea8c1571df544cb6a8d112e57029fb42054d27c54146\": container with ID starting with 5086f2ed6a5667b1e238ea8c1571df544cb6a8d112e57029fb42054d27c54146 not found: ID does not exist" containerID="5086f2ed6a5667b1e238ea8c1571df544cb6a8d112e57029fb42054d27c54146" Oct 01 09:59:25 crc kubenswrapper[4787]: I1001 09:59:25.849430 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5086f2ed6a5667b1e238ea8c1571df544cb6a8d112e57029fb42054d27c54146"} err="failed to get container status \"5086f2ed6a5667b1e238ea8c1571df544cb6a8d112e57029fb42054d27c54146\": rpc error: code = NotFound desc = could not find container \"5086f2ed6a5667b1e238ea8c1571df544cb6a8d112e57029fb42054d27c54146\": container with ID starting with 5086f2ed6a5667b1e238ea8c1571df544cb6a8d112e57029fb42054d27c54146 not found: ID does not exist" Oct 01 09:59:26 crc kubenswrapper[4787]: I1001 09:59:26.553610 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1af17f67-b247-48f9-a000-83d626489e29" path="/var/lib/kubelet/pods/1af17f67-b247-48f9-a000-83d626489e29/volumes" Oct 01 09:59:30 crc kubenswrapper[4787]: I1001 09:59:30.287481 4787 scope.go:117] "RemoveContainer" containerID="cfb994f04e37d89a8d2e9681493c188d01b7b7dad216fc2587b4419c5a95e425" Oct 01 09:59:30 crc kubenswrapper[4787]: I1001 09:59:30.308142 4787 scope.go:117] "RemoveContainer" containerID="f604cceb35751cdf4ced9c10d851eef06a4de09e5f2c818aa1f04a04d70ef53c" Oct 01 09:59:31 crc kubenswrapper[4787]: I1001 09:59:31.447661 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-s4vj6"] Oct 01 09:59:31 crc kubenswrapper[4787]: E1001 09:59:31.452166 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1af17f67-b247-48f9-a000-83d626489e29" containerName="extract-utilities" Oct 01 09:59:31 crc kubenswrapper[4787]: I1001 09:59:31.452203 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="1af17f67-b247-48f9-a000-83d626489e29" containerName="extract-utilities" Oct 01 09:59:31 crc kubenswrapper[4787]: E1001 09:59:31.452242 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1af17f67-b247-48f9-a000-83d626489e29" containerName="registry-server" Oct 01 09:59:31 crc kubenswrapper[4787]: I1001 09:59:31.452252 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="1af17f67-b247-48f9-a000-83d626489e29" containerName="registry-server" Oct 01 09:59:31 crc kubenswrapper[4787]: E1001 09:59:31.452298 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1af17f67-b247-48f9-a000-83d626489e29" containerName="extract-content" Oct 01 09:59:31 crc kubenswrapper[4787]: I1001 09:59:31.452307 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="1af17f67-b247-48f9-a000-83d626489e29" containerName="extract-content" Oct 01 09:59:31 crc kubenswrapper[4787]: I1001 09:59:31.453012 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="1af17f67-b247-48f9-a000-83d626489e29" containerName="registry-server" Oct 01 09:59:31 crc kubenswrapper[4787]: I1001 09:59:31.456861 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s4vj6" Oct 01 09:59:31 crc kubenswrapper[4787]: I1001 09:59:31.469619 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-s4vj6"] Oct 01 09:59:31 crc kubenswrapper[4787]: I1001 09:59:31.561249 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f3cb990-6f7c-45d6-a061-1b4a2afe3340-catalog-content\") pod \"redhat-operators-s4vj6\" (UID: \"2f3cb990-6f7c-45d6-a061-1b4a2afe3340\") " pod="openshift-marketplace/redhat-operators-s4vj6" Oct 01 09:59:31 crc kubenswrapper[4787]: I1001 09:59:31.561413 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f3cb990-6f7c-45d6-a061-1b4a2afe3340-utilities\") pod \"redhat-operators-s4vj6\" (UID: \"2f3cb990-6f7c-45d6-a061-1b4a2afe3340\") " pod="openshift-marketplace/redhat-operators-s4vj6" Oct 01 09:59:31 crc kubenswrapper[4787]: I1001 09:59:31.561662 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svfx9\" (UniqueName: \"kubernetes.io/projected/2f3cb990-6f7c-45d6-a061-1b4a2afe3340-kube-api-access-svfx9\") pod \"redhat-operators-s4vj6\" (UID: \"2f3cb990-6f7c-45d6-a061-1b4a2afe3340\") " pod="openshift-marketplace/redhat-operators-s4vj6" Oct 01 09:59:31 crc kubenswrapper[4787]: I1001 09:59:31.663779 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f3cb990-6f7c-45d6-a061-1b4a2afe3340-catalog-content\") pod \"redhat-operators-s4vj6\" (UID: \"2f3cb990-6f7c-45d6-a061-1b4a2afe3340\") " pod="openshift-marketplace/redhat-operators-s4vj6" Oct 01 09:59:31 crc kubenswrapper[4787]: I1001 09:59:31.663862 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f3cb990-6f7c-45d6-a061-1b4a2afe3340-utilities\") pod \"redhat-operators-s4vj6\" (UID: \"2f3cb990-6f7c-45d6-a061-1b4a2afe3340\") " pod="openshift-marketplace/redhat-operators-s4vj6" Oct 01 09:59:31 crc kubenswrapper[4787]: I1001 09:59:31.663938 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svfx9\" (UniqueName: \"kubernetes.io/projected/2f3cb990-6f7c-45d6-a061-1b4a2afe3340-kube-api-access-svfx9\") pod \"redhat-operators-s4vj6\" (UID: \"2f3cb990-6f7c-45d6-a061-1b4a2afe3340\") " pod="openshift-marketplace/redhat-operators-s4vj6" Oct 01 09:59:31 crc kubenswrapper[4787]: I1001 09:59:31.664377 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f3cb990-6f7c-45d6-a061-1b4a2afe3340-catalog-content\") pod \"redhat-operators-s4vj6\" (UID: \"2f3cb990-6f7c-45d6-a061-1b4a2afe3340\") " pod="openshift-marketplace/redhat-operators-s4vj6" Oct 01 09:59:31 crc kubenswrapper[4787]: I1001 09:59:31.664480 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f3cb990-6f7c-45d6-a061-1b4a2afe3340-utilities\") pod \"redhat-operators-s4vj6\" (UID: \"2f3cb990-6f7c-45d6-a061-1b4a2afe3340\") " pod="openshift-marketplace/redhat-operators-s4vj6" Oct 01 09:59:31 crc kubenswrapper[4787]: I1001 09:59:31.688070 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svfx9\" (UniqueName: \"kubernetes.io/projected/2f3cb990-6f7c-45d6-a061-1b4a2afe3340-kube-api-access-svfx9\") pod \"redhat-operators-s4vj6\" (UID: \"2f3cb990-6f7c-45d6-a061-1b4a2afe3340\") " pod="openshift-marketplace/redhat-operators-s4vj6" Oct 01 09:59:31 crc kubenswrapper[4787]: I1001 09:59:31.790405 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s4vj6" Oct 01 09:59:32 crc kubenswrapper[4787]: I1001 09:59:32.300992 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-s4vj6"] Oct 01 09:59:32 crc kubenswrapper[4787]: I1001 09:59:32.826736 4787 generic.go:334] "Generic (PLEG): container finished" podID="2f3cb990-6f7c-45d6-a061-1b4a2afe3340" containerID="4224ae9316f015b60bcec7fa0071bbe3f05a2239bac5ea069ef9cb9ad4de1df8" exitCode=0 Oct 01 09:59:32 crc kubenswrapper[4787]: I1001 09:59:32.827713 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s4vj6" event={"ID":"2f3cb990-6f7c-45d6-a061-1b4a2afe3340","Type":"ContainerDied","Data":"4224ae9316f015b60bcec7fa0071bbe3f05a2239bac5ea069ef9cb9ad4de1df8"} Oct 01 09:59:32 crc kubenswrapper[4787]: I1001 09:59:32.827750 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s4vj6" event={"ID":"2f3cb990-6f7c-45d6-a061-1b4a2afe3340","Type":"ContainerStarted","Data":"080041f0ee9d736fe1112b97135d0214038813e7ec34aea06401dc888fd2cbd8"} Oct 01 09:59:34 crc kubenswrapper[4787]: I1001 09:59:34.673769 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zwsdr" Oct 01 09:59:34 crc kubenswrapper[4787]: E1001 09:59:34.735789 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2f3cb990_6f7c_45d6_a061_1b4a2afe3340.slice/crio-df4c8f52ca6a8599b50827f0ba5af46edcb0d48435e089c064b95d9ed980a882.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2f3cb990_6f7c_45d6_a061_1b4a2afe3340.slice/crio-conmon-df4c8f52ca6a8599b50827f0ba5af46edcb0d48435e089c064b95d9ed980a882.scope\": RecentStats: unable to find data in memory cache]" Oct 01 09:59:34 crc kubenswrapper[4787]: I1001 09:59:34.852895 4787 generic.go:334] "Generic (PLEG): container finished" podID="2f3cb990-6f7c-45d6-a061-1b4a2afe3340" containerID="df4c8f52ca6a8599b50827f0ba5af46edcb0d48435e089c064b95d9ed980a882" exitCode=0 Oct 01 09:59:34 crc kubenswrapper[4787]: I1001 09:59:34.852988 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s4vj6" event={"ID":"2f3cb990-6f7c-45d6-a061-1b4a2afe3340","Type":"ContainerDied","Data":"df4c8f52ca6a8599b50827f0ba5af46edcb0d48435e089c064b95d9ed980a882"} Oct 01 09:59:35 crc kubenswrapper[4787]: I1001 09:59:35.874062 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s4vj6" event={"ID":"2f3cb990-6f7c-45d6-a061-1b4a2afe3340","Type":"ContainerStarted","Data":"bb748880d10a1eff67d89e957973c373ee121f232e19469112f6723a2108aca1"} Oct 01 09:59:35 crc kubenswrapper[4787]: I1001 09:59:35.903824 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-s4vj6" podStartSLOduration=2.429702763 podStartE2EDuration="4.903801653s" podCreationTimestamp="2025-10-01 09:59:31 +0000 UTC" firstStartedPulling="2025-10-01 09:59:32.830324966 +0000 UTC m=+1404.945469123" lastFinishedPulling="2025-10-01 09:59:35.304423816 +0000 UTC m=+1407.419568013" observedRunningTime="2025-10-01 09:59:35.902653555 +0000 UTC m=+1408.017797712" watchObservedRunningTime="2025-10-01 09:59:35.903801653 +0000 UTC m=+1408.018945810" Oct 01 09:59:36 crc kubenswrapper[4787]: I1001 09:59:36.597399 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zwsdr"] Oct 01 09:59:36 crc kubenswrapper[4787]: I1001 09:59:36.597693 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zwsdr" podUID="6d846f41-b98b-41fd-b027-fd764c01c021" containerName="registry-server" containerID="cri-o://a217e19f0304c1f5b923a84decadd1d7d323ad112afd080d16c4e377bed351e5" gracePeriod=2 Oct 01 09:59:36 crc kubenswrapper[4787]: I1001 09:59:36.887296 4787 generic.go:334] "Generic (PLEG): container finished" podID="6d846f41-b98b-41fd-b027-fd764c01c021" containerID="a217e19f0304c1f5b923a84decadd1d7d323ad112afd080d16c4e377bed351e5" exitCode=0 Oct 01 09:59:36 crc kubenswrapper[4787]: I1001 09:59:36.887517 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zwsdr" event={"ID":"6d846f41-b98b-41fd-b027-fd764c01c021","Type":"ContainerDied","Data":"a217e19f0304c1f5b923a84decadd1d7d323ad112afd080d16c4e377bed351e5"} Oct 01 09:59:37 crc kubenswrapper[4787]: I1001 09:59:37.588491 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zwsdr" Oct 01 09:59:37 crc kubenswrapper[4787]: I1001 09:59:37.692002 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d846f41-b98b-41fd-b027-fd764c01c021-catalog-content\") pod \"6d846f41-b98b-41fd-b027-fd764c01c021\" (UID: \"6d846f41-b98b-41fd-b027-fd764c01c021\") " Oct 01 09:59:37 crc kubenswrapper[4787]: I1001 09:59:37.693278 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d846f41-b98b-41fd-b027-fd764c01c021-utilities\") pod \"6d846f41-b98b-41fd-b027-fd764c01c021\" (UID: \"6d846f41-b98b-41fd-b027-fd764c01c021\") " Oct 01 09:59:37 crc kubenswrapper[4787]: I1001 09:59:37.693355 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77mdj\" (UniqueName: \"kubernetes.io/projected/6d846f41-b98b-41fd-b027-fd764c01c021-kube-api-access-77mdj\") pod \"6d846f41-b98b-41fd-b027-fd764c01c021\" (UID: \"6d846f41-b98b-41fd-b027-fd764c01c021\") " Oct 01 09:59:37 crc kubenswrapper[4787]: I1001 09:59:37.693920 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d846f41-b98b-41fd-b027-fd764c01c021-utilities" (OuterVolumeSpecName: "utilities") pod "6d846f41-b98b-41fd-b027-fd764c01c021" (UID: "6d846f41-b98b-41fd-b027-fd764c01c021"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:59:37 crc kubenswrapper[4787]: I1001 09:59:37.702784 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d846f41-b98b-41fd-b027-fd764c01c021-kube-api-access-77mdj" (OuterVolumeSpecName: "kube-api-access-77mdj") pod "6d846f41-b98b-41fd-b027-fd764c01c021" (UID: "6d846f41-b98b-41fd-b027-fd764c01c021"). InnerVolumeSpecName "kube-api-access-77mdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:59:37 crc kubenswrapper[4787]: I1001 09:59:37.734364 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d846f41-b98b-41fd-b027-fd764c01c021-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6d846f41-b98b-41fd-b027-fd764c01c021" (UID: "6d846f41-b98b-41fd-b027-fd764c01c021"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:59:37 crc kubenswrapper[4787]: I1001 09:59:37.795413 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d846f41-b98b-41fd-b027-fd764c01c021-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:59:37 crc kubenswrapper[4787]: I1001 09:59:37.795452 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d846f41-b98b-41fd-b027-fd764c01c021-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:59:37 crc kubenswrapper[4787]: I1001 09:59:37.795467 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77mdj\" (UniqueName: \"kubernetes.io/projected/6d846f41-b98b-41fd-b027-fd764c01c021-kube-api-access-77mdj\") on node \"crc\" DevicePath \"\"" Oct 01 09:59:37 crc kubenswrapper[4787]: I1001 09:59:37.898588 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zwsdr" event={"ID":"6d846f41-b98b-41fd-b027-fd764c01c021","Type":"ContainerDied","Data":"ee439afd56d3c0674acdb73c29821a6d08db687228b3b76a8545615239a69713"} Oct 01 09:59:37 crc kubenswrapper[4787]: I1001 09:59:37.898672 4787 scope.go:117] "RemoveContainer" containerID="a217e19f0304c1f5b923a84decadd1d7d323ad112afd080d16c4e377bed351e5" Oct 01 09:59:37 crc kubenswrapper[4787]: I1001 09:59:37.898656 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zwsdr" Oct 01 09:59:37 crc kubenswrapper[4787]: I1001 09:59:37.928563 4787 scope.go:117] "RemoveContainer" containerID="0ad5444728a5182df188d71b14576784ac47c71ee00c4a769c0b55c9d9409c56" Oct 01 09:59:37 crc kubenswrapper[4787]: I1001 09:59:37.942705 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zwsdr"] Oct 01 09:59:37 crc kubenswrapper[4787]: I1001 09:59:37.956853 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zwsdr"] Oct 01 09:59:37 crc kubenswrapper[4787]: I1001 09:59:37.961786 4787 scope.go:117] "RemoveContainer" containerID="810d10621dfc5cf0561dc2847098f182f30774f720d46ed5ae306e1f86da5b51" Oct 01 09:59:38 crc kubenswrapper[4787]: I1001 09:59:38.535858 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d846f41-b98b-41fd-b027-fd764c01c021" path="/var/lib/kubelet/pods/6d846f41-b98b-41fd-b027-fd764c01c021/volumes" Oct 01 09:59:41 crc kubenswrapper[4787]: I1001 09:59:41.251187 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 09:59:41 crc kubenswrapper[4787]: I1001 09:59:41.251289 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 09:59:41 crc kubenswrapper[4787]: I1001 09:59:41.791121 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-s4vj6" Oct 01 09:59:41 crc kubenswrapper[4787]: I1001 09:59:41.791185 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-s4vj6" Oct 01 09:59:41 crc kubenswrapper[4787]: I1001 09:59:41.939748 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-s4vj6" Oct 01 09:59:41 crc kubenswrapper[4787]: I1001 09:59:41.985460 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-s4vj6" Oct 01 09:59:43 crc kubenswrapper[4787]: I1001 09:59:43.002980 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-s4vj6"] Oct 01 09:59:43 crc kubenswrapper[4787]: I1001 09:59:43.958018 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-s4vj6" podUID="2f3cb990-6f7c-45d6-a061-1b4a2afe3340" containerName="registry-server" containerID="cri-o://bb748880d10a1eff67d89e957973c373ee121f232e19469112f6723a2108aca1" gracePeriod=2 Oct 01 09:59:44 crc kubenswrapper[4787]: I1001 09:59:44.406984 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s4vj6" Oct 01 09:59:44 crc kubenswrapper[4787]: I1001 09:59:44.541513 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svfx9\" (UniqueName: \"kubernetes.io/projected/2f3cb990-6f7c-45d6-a061-1b4a2afe3340-kube-api-access-svfx9\") pod \"2f3cb990-6f7c-45d6-a061-1b4a2afe3340\" (UID: \"2f3cb990-6f7c-45d6-a061-1b4a2afe3340\") " Oct 01 09:59:44 crc kubenswrapper[4787]: I1001 09:59:44.541760 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f3cb990-6f7c-45d6-a061-1b4a2afe3340-catalog-content\") pod \"2f3cb990-6f7c-45d6-a061-1b4a2afe3340\" (UID: \"2f3cb990-6f7c-45d6-a061-1b4a2afe3340\") " Oct 01 09:59:44 crc kubenswrapper[4787]: I1001 09:59:44.541841 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f3cb990-6f7c-45d6-a061-1b4a2afe3340-utilities\") pod \"2f3cb990-6f7c-45d6-a061-1b4a2afe3340\" (UID: \"2f3cb990-6f7c-45d6-a061-1b4a2afe3340\") " Oct 01 09:59:44 crc kubenswrapper[4787]: I1001 09:59:44.542797 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f3cb990-6f7c-45d6-a061-1b4a2afe3340-utilities" (OuterVolumeSpecName: "utilities") pod "2f3cb990-6f7c-45d6-a061-1b4a2afe3340" (UID: "2f3cb990-6f7c-45d6-a061-1b4a2afe3340"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:59:44 crc kubenswrapper[4787]: I1001 09:59:44.561458 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f3cb990-6f7c-45d6-a061-1b4a2afe3340-kube-api-access-svfx9" (OuterVolumeSpecName: "kube-api-access-svfx9") pod "2f3cb990-6f7c-45d6-a061-1b4a2afe3340" (UID: "2f3cb990-6f7c-45d6-a061-1b4a2afe3340"). InnerVolumeSpecName "kube-api-access-svfx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:59:44 crc kubenswrapper[4787]: I1001 09:59:44.632971 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f3cb990-6f7c-45d6-a061-1b4a2afe3340-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2f3cb990-6f7c-45d6-a061-1b4a2afe3340" (UID: "2f3cb990-6f7c-45d6-a061-1b4a2afe3340"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 09:59:44 crc kubenswrapper[4787]: I1001 09:59:44.644267 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f3cb990-6f7c-45d6-a061-1b4a2afe3340-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 09:59:44 crc kubenswrapper[4787]: I1001 09:59:44.644304 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f3cb990-6f7c-45d6-a061-1b4a2afe3340-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 09:59:44 crc kubenswrapper[4787]: I1001 09:59:44.644318 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svfx9\" (UniqueName: \"kubernetes.io/projected/2f3cb990-6f7c-45d6-a061-1b4a2afe3340-kube-api-access-svfx9\") on node \"crc\" DevicePath \"\"" Oct 01 09:59:44 crc kubenswrapper[4787]: I1001 09:59:44.985041 4787 generic.go:334] "Generic (PLEG): container finished" podID="2f3cb990-6f7c-45d6-a061-1b4a2afe3340" containerID="bb748880d10a1eff67d89e957973c373ee121f232e19469112f6723a2108aca1" exitCode=0 Oct 01 09:59:44 crc kubenswrapper[4787]: I1001 09:59:44.985535 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s4vj6" event={"ID":"2f3cb990-6f7c-45d6-a061-1b4a2afe3340","Type":"ContainerDied","Data":"bb748880d10a1eff67d89e957973c373ee121f232e19469112f6723a2108aca1"} Oct 01 09:59:44 crc kubenswrapper[4787]: I1001 09:59:44.985572 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s4vj6" event={"ID":"2f3cb990-6f7c-45d6-a061-1b4a2afe3340","Type":"ContainerDied","Data":"080041f0ee9d736fe1112b97135d0214038813e7ec34aea06401dc888fd2cbd8"} Oct 01 09:59:44 crc kubenswrapper[4787]: I1001 09:59:44.985595 4787 scope.go:117] "RemoveContainer" containerID="bb748880d10a1eff67d89e957973c373ee121f232e19469112f6723a2108aca1" Oct 01 09:59:44 crc kubenswrapper[4787]: I1001 09:59:44.985769 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s4vj6" Oct 01 09:59:45 crc kubenswrapper[4787]: I1001 09:59:45.013409 4787 scope.go:117] "RemoveContainer" containerID="df4c8f52ca6a8599b50827f0ba5af46edcb0d48435e089c064b95d9ed980a882" Oct 01 09:59:45 crc kubenswrapper[4787]: I1001 09:59:45.042497 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-s4vj6"] Oct 01 09:59:45 crc kubenswrapper[4787]: I1001 09:59:45.050000 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-s4vj6"] Oct 01 09:59:45 crc kubenswrapper[4787]: I1001 09:59:45.052511 4787 scope.go:117] "RemoveContainer" containerID="4224ae9316f015b60bcec7fa0071bbe3f05a2239bac5ea069ef9cb9ad4de1df8" Oct 01 09:59:45 crc kubenswrapper[4787]: I1001 09:59:45.075623 4787 scope.go:117] "RemoveContainer" containerID="bb748880d10a1eff67d89e957973c373ee121f232e19469112f6723a2108aca1" Oct 01 09:59:45 crc kubenswrapper[4787]: E1001 09:59:45.076068 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb748880d10a1eff67d89e957973c373ee121f232e19469112f6723a2108aca1\": container with ID starting with bb748880d10a1eff67d89e957973c373ee121f232e19469112f6723a2108aca1 not found: ID does not exist" containerID="bb748880d10a1eff67d89e957973c373ee121f232e19469112f6723a2108aca1" Oct 01 09:59:45 crc kubenswrapper[4787]: I1001 09:59:45.076147 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb748880d10a1eff67d89e957973c373ee121f232e19469112f6723a2108aca1"} err="failed to get container status \"bb748880d10a1eff67d89e957973c373ee121f232e19469112f6723a2108aca1\": rpc error: code = NotFound desc = could not find container \"bb748880d10a1eff67d89e957973c373ee121f232e19469112f6723a2108aca1\": container with ID starting with bb748880d10a1eff67d89e957973c373ee121f232e19469112f6723a2108aca1 not found: ID does not exist" Oct 01 09:59:45 crc kubenswrapper[4787]: I1001 09:59:45.076176 4787 scope.go:117] "RemoveContainer" containerID="df4c8f52ca6a8599b50827f0ba5af46edcb0d48435e089c064b95d9ed980a882" Oct 01 09:59:45 crc kubenswrapper[4787]: E1001 09:59:45.076574 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df4c8f52ca6a8599b50827f0ba5af46edcb0d48435e089c064b95d9ed980a882\": container with ID starting with df4c8f52ca6a8599b50827f0ba5af46edcb0d48435e089c064b95d9ed980a882 not found: ID does not exist" containerID="df4c8f52ca6a8599b50827f0ba5af46edcb0d48435e089c064b95d9ed980a882" Oct 01 09:59:45 crc kubenswrapper[4787]: I1001 09:59:45.076634 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df4c8f52ca6a8599b50827f0ba5af46edcb0d48435e089c064b95d9ed980a882"} err="failed to get container status \"df4c8f52ca6a8599b50827f0ba5af46edcb0d48435e089c064b95d9ed980a882\": rpc error: code = NotFound desc = could not find container \"df4c8f52ca6a8599b50827f0ba5af46edcb0d48435e089c064b95d9ed980a882\": container with ID starting with df4c8f52ca6a8599b50827f0ba5af46edcb0d48435e089c064b95d9ed980a882 not found: ID does not exist" Oct 01 09:59:45 crc kubenswrapper[4787]: I1001 09:59:45.076668 4787 scope.go:117] "RemoveContainer" containerID="4224ae9316f015b60bcec7fa0071bbe3f05a2239bac5ea069ef9cb9ad4de1df8" Oct 01 09:59:45 crc kubenswrapper[4787]: E1001 09:59:45.076890 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4224ae9316f015b60bcec7fa0071bbe3f05a2239bac5ea069ef9cb9ad4de1df8\": container with ID starting with 4224ae9316f015b60bcec7fa0071bbe3f05a2239bac5ea069ef9cb9ad4de1df8 not found: ID does not exist" containerID="4224ae9316f015b60bcec7fa0071bbe3f05a2239bac5ea069ef9cb9ad4de1df8" Oct 01 09:59:45 crc kubenswrapper[4787]: I1001 09:59:45.076911 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4224ae9316f015b60bcec7fa0071bbe3f05a2239bac5ea069ef9cb9ad4de1df8"} err="failed to get container status \"4224ae9316f015b60bcec7fa0071bbe3f05a2239bac5ea069ef9cb9ad4de1df8\": rpc error: code = NotFound desc = could not find container \"4224ae9316f015b60bcec7fa0071bbe3f05a2239bac5ea069ef9cb9ad4de1df8\": container with ID starting with 4224ae9316f015b60bcec7fa0071bbe3f05a2239bac5ea069ef9cb9ad4de1df8 not found: ID does not exist" Oct 01 09:59:46 crc kubenswrapper[4787]: I1001 09:59:46.537631 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f3cb990-6f7c-45d6-a061-1b4a2afe3340" path="/var/lib/kubelet/pods/2f3cb990-6f7c-45d6-a061-1b4a2afe3340/volumes" Oct 01 09:59:54 crc kubenswrapper[4787]: I1001 09:59:54.083821 4787 generic.go:334] "Generic (PLEG): container finished" podID="d1018ece-ee61-4913-bd2f-64f996ef135c" containerID="b8edd8c7dc76be74e1f30784b4e406d40f803fd07cbb20f81682e17580421ed7" exitCode=0 Oct 01 09:59:54 crc kubenswrapper[4787]: I1001 09:59:54.083915 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s4tf6" event={"ID":"d1018ece-ee61-4913-bd2f-64f996ef135c","Type":"ContainerDied","Data":"b8edd8c7dc76be74e1f30784b4e406d40f803fd07cbb20f81682e17580421ed7"} Oct 01 09:59:55 crc kubenswrapper[4787]: I1001 09:59:55.568781 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s4tf6" Oct 01 09:59:55 crc kubenswrapper[4787]: I1001 09:59:55.739756 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6h6dv\" (UniqueName: \"kubernetes.io/projected/d1018ece-ee61-4913-bd2f-64f996ef135c-kube-api-access-6h6dv\") pod \"d1018ece-ee61-4913-bd2f-64f996ef135c\" (UID: \"d1018ece-ee61-4913-bd2f-64f996ef135c\") " Oct 01 09:59:55 crc kubenswrapper[4787]: I1001 09:59:55.739903 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1018ece-ee61-4913-bd2f-64f996ef135c-bootstrap-combined-ca-bundle\") pod \"d1018ece-ee61-4913-bd2f-64f996ef135c\" (UID: \"d1018ece-ee61-4913-bd2f-64f996ef135c\") " Oct 01 09:59:55 crc kubenswrapper[4787]: I1001 09:59:55.739968 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d1018ece-ee61-4913-bd2f-64f996ef135c-inventory\") pod \"d1018ece-ee61-4913-bd2f-64f996ef135c\" (UID: \"d1018ece-ee61-4913-bd2f-64f996ef135c\") " Oct 01 09:59:55 crc kubenswrapper[4787]: I1001 09:59:55.739985 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d1018ece-ee61-4913-bd2f-64f996ef135c-ssh-key\") pod \"d1018ece-ee61-4913-bd2f-64f996ef135c\" (UID: \"d1018ece-ee61-4913-bd2f-64f996ef135c\") " Oct 01 09:59:55 crc kubenswrapper[4787]: I1001 09:59:55.747967 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1018ece-ee61-4913-bd2f-64f996ef135c-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "d1018ece-ee61-4913-bd2f-64f996ef135c" (UID: "d1018ece-ee61-4913-bd2f-64f996ef135c"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:59:55 crc kubenswrapper[4787]: I1001 09:59:55.748036 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1018ece-ee61-4913-bd2f-64f996ef135c-kube-api-access-6h6dv" (OuterVolumeSpecName: "kube-api-access-6h6dv") pod "d1018ece-ee61-4913-bd2f-64f996ef135c" (UID: "d1018ece-ee61-4913-bd2f-64f996ef135c"). InnerVolumeSpecName "kube-api-access-6h6dv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 09:59:55 crc kubenswrapper[4787]: I1001 09:59:55.778552 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1018ece-ee61-4913-bd2f-64f996ef135c-inventory" (OuterVolumeSpecName: "inventory") pod "d1018ece-ee61-4913-bd2f-64f996ef135c" (UID: "d1018ece-ee61-4913-bd2f-64f996ef135c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:59:55 crc kubenswrapper[4787]: I1001 09:59:55.781002 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1018ece-ee61-4913-bd2f-64f996ef135c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d1018ece-ee61-4913-bd2f-64f996ef135c" (UID: "d1018ece-ee61-4913-bd2f-64f996ef135c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 09:59:55 crc kubenswrapper[4787]: I1001 09:59:55.842399 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6h6dv\" (UniqueName: \"kubernetes.io/projected/d1018ece-ee61-4913-bd2f-64f996ef135c-kube-api-access-6h6dv\") on node \"crc\" DevicePath \"\"" Oct 01 09:59:55 crc kubenswrapper[4787]: I1001 09:59:55.842494 4787 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1018ece-ee61-4913-bd2f-64f996ef135c-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 09:59:55 crc kubenswrapper[4787]: I1001 09:59:55.842517 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d1018ece-ee61-4913-bd2f-64f996ef135c-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 09:59:55 crc kubenswrapper[4787]: I1001 09:59:55.842532 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d1018ece-ee61-4913-bd2f-64f996ef135c-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 09:59:56 crc kubenswrapper[4787]: I1001 09:59:56.121229 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s4tf6" event={"ID":"d1018ece-ee61-4913-bd2f-64f996ef135c","Type":"ContainerDied","Data":"1604dcca1e397a31c19754c01dd5f17702126d7b51b04c38aa226d4e07e39103"} Oct 01 09:59:56 crc kubenswrapper[4787]: I1001 09:59:56.121280 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1604dcca1e397a31c19754c01dd5f17702126d7b51b04c38aa226d4e07e39103" Oct 01 09:59:56 crc kubenswrapper[4787]: I1001 09:59:56.121432 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s4tf6" Oct 01 09:59:56 crc kubenswrapper[4787]: I1001 09:59:56.220565 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4jp68"] Oct 01 09:59:56 crc kubenswrapper[4787]: E1001 09:59:56.221228 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d846f41-b98b-41fd-b027-fd764c01c021" containerName="registry-server" Oct 01 09:59:56 crc kubenswrapper[4787]: I1001 09:59:56.221252 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d846f41-b98b-41fd-b027-fd764c01c021" containerName="registry-server" Oct 01 09:59:56 crc kubenswrapper[4787]: E1001 09:59:56.221289 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d846f41-b98b-41fd-b027-fd764c01c021" containerName="extract-content" Oct 01 09:59:56 crc kubenswrapper[4787]: I1001 09:59:56.221298 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d846f41-b98b-41fd-b027-fd764c01c021" containerName="extract-content" Oct 01 09:59:56 crc kubenswrapper[4787]: E1001 09:59:56.221314 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f3cb990-6f7c-45d6-a061-1b4a2afe3340" containerName="registry-server" Oct 01 09:59:56 crc kubenswrapper[4787]: I1001 09:59:56.221326 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f3cb990-6f7c-45d6-a061-1b4a2afe3340" containerName="registry-server" Oct 01 09:59:56 crc kubenswrapper[4787]: E1001 09:59:56.221342 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f3cb990-6f7c-45d6-a061-1b4a2afe3340" containerName="extract-utilities" Oct 01 09:59:56 crc kubenswrapper[4787]: I1001 09:59:56.221352 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f3cb990-6f7c-45d6-a061-1b4a2afe3340" containerName="extract-utilities" Oct 01 09:59:56 crc kubenswrapper[4787]: E1001 09:59:56.221375 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1018ece-ee61-4913-bd2f-64f996ef135c" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 01 09:59:56 crc kubenswrapper[4787]: I1001 09:59:56.221386 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1018ece-ee61-4913-bd2f-64f996ef135c" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 01 09:59:56 crc kubenswrapper[4787]: E1001 09:59:56.221402 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d846f41-b98b-41fd-b027-fd764c01c021" containerName="extract-utilities" Oct 01 09:59:56 crc kubenswrapper[4787]: I1001 09:59:56.221411 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d846f41-b98b-41fd-b027-fd764c01c021" containerName="extract-utilities" Oct 01 09:59:56 crc kubenswrapper[4787]: E1001 09:59:56.221431 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f3cb990-6f7c-45d6-a061-1b4a2afe3340" containerName="extract-content" Oct 01 09:59:56 crc kubenswrapper[4787]: I1001 09:59:56.221442 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f3cb990-6f7c-45d6-a061-1b4a2afe3340" containerName="extract-content" Oct 01 09:59:56 crc kubenswrapper[4787]: I1001 09:59:56.221714 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1018ece-ee61-4913-bd2f-64f996ef135c" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 01 09:59:56 crc kubenswrapper[4787]: I1001 09:59:56.221730 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d846f41-b98b-41fd-b027-fd764c01c021" containerName="registry-server" Oct 01 09:59:56 crc kubenswrapper[4787]: I1001 09:59:56.221761 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f3cb990-6f7c-45d6-a061-1b4a2afe3340" containerName="registry-server" Oct 01 09:59:56 crc kubenswrapper[4787]: I1001 09:59:56.222852 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4jp68" Oct 01 09:59:56 crc kubenswrapper[4787]: I1001 09:59:56.230110 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vh5v6" Oct 01 09:59:56 crc kubenswrapper[4787]: I1001 09:59:56.230467 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 09:59:56 crc kubenswrapper[4787]: I1001 09:59:56.230467 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 09:59:56 crc kubenswrapper[4787]: I1001 09:59:56.232901 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 09:59:56 crc kubenswrapper[4787]: I1001 09:59:56.236921 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4jp68"] Oct 01 09:59:56 crc kubenswrapper[4787]: I1001 09:59:56.353377 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58tln\" (UniqueName: \"kubernetes.io/projected/f70ffa29-bf25-44a2-bd45-822e5f59a5d5-kube-api-access-58tln\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4jp68\" (UID: \"f70ffa29-bf25-44a2-bd45-822e5f59a5d5\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4jp68" Oct 01 09:59:56 crc kubenswrapper[4787]: I1001 09:59:56.353590 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f70ffa29-bf25-44a2-bd45-822e5f59a5d5-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4jp68\" (UID: \"f70ffa29-bf25-44a2-bd45-822e5f59a5d5\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4jp68" Oct 01 09:59:56 crc kubenswrapper[4787]: I1001 09:59:56.353906 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f70ffa29-bf25-44a2-bd45-822e5f59a5d5-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4jp68\" (UID: \"f70ffa29-bf25-44a2-bd45-822e5f59a5d5\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4jp68" Oct 01 09:59:56 crc kubenswrapper[4787]: I1001 09:59:56.455953 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f70ffa29-bf25-44a2-bd45-822e5f59a5d5-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4jp68\" (UID: \"f70ffa29-bf25-44a2-bd45-822e5f59a5d5\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4jp68" Oct 01 09:59:56 crc kubenswrapper[4787]: I1001 09:59:56.456170 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58tln\" (UniqueName: \"kubernetes.io/projected/f70ffa29-bf25-44a2-bd45-822e5f59a5d5-kube-api-access-58tln\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4jp68\" (UID: \"f70ffa29-bf25-44a2-bd45-822e5f59a5d5\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4jp68" Oct 01 09:59:56 crc kubenswrapper[4787]: I1001 09:59:56.456216 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f70ffa29-bf25-44a2-bd45-822e5f59a5d5-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4jp68\" (UID: \"f70ffa29-bf25-44a2-bd45-822e5f59a5d5\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4jp68" Oct 01 09:59:56 crc kubenswrapper[4787]: I1001 09:59:56.460030 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f70ffa29-bf25-44a2-bd45-822e5f59a5d5-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4jp68\" (UID: \"f70ffa29-bf25-44a2-bd45-822e5f59a5d5\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4jp68" Oct 01 09:59:56 crc kubenswrapper[4787]: I1001 09:59:56.460994 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f70ffa29-bf25-44a2-bd45-822e5f59a5d5-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4jp68\" (UID: \"f70ffa29-bf25-44a2-bd45-822e5f59a5d5\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4jp68" Oct 01 09:59:56 crc kubenswrapper[4787]: I1001 09:59:56.478273 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58tln\" (UniqueName: \"kubernetes.io/projected/f70ffa29-bf25-44a2-bd45-822e5f59a5d5-kube-api-access-58tln\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4jp68\" (UID: \"f70ffa29-bf25-44a2-bd45-822e5f59a5d5\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4jp68" Oct 01 09:59:56 crc kubenswrapper[4787]: I1001 09:59:56.560840 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4jp68" Oct 01 09:59:57 crc kubenswrapper[4787]: I1001 09:59:57.088876 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4jp68"] Oct 01 09:59:57 crc kubenswrapper[4787]: I1001 09:59:57.134932 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4jp68" event={"ID":"f70ffa29-bf25-44a2-bd45-822e5f59a5d5","Type":"ContainerStarted","Data":"33d40988c9c35849c4c440ced9ca6bbad3802d44bbadfe74266adfc65ec8eda7"} Oct 01 09:59:58 crc kubenswrapper[4787]: I1001 09:59:58.146437 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4jp68" event={"ID":"f70ffa29-bf25-44a2-bd45-822e5f59a5d5","Type":"ContainerStarted","Data":"7cf3e48f00ff3dc03265b85bf5517015639767c21f22edd24096e676468b2266"} Oct 01 09:59:58 crc kubenswrapper[4787]: I1001 09:59:58.175510 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4jp68" podStartSLOduration=1.7093072249999999 podStartE2EDuration="2.175483269s" podCreationTimestamp="2025-10-01 09:59:56 +0000 UTC" firstStartedPulling="2025-10-01 09:59:57.095543178 +0000 UTC m=+1429.210687375" lastFinishedPulling="2025-10-01 09:59:57.561719222 +0000 UTC m=+1429.676863419" observedRunningTime="2025-10-01 09:59:58.16686183 +0000 UTC m=+1430.282006077" watchObservedRunningTime="2025-10-01 09:59:58.175483269 +0000 UTC m=+1430.290627456" Oct 01 10:00:00 crc kubenswrapper[4787]: I1001 10:00:00.132991 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321880-mxzcx"] Oct 01 10:00:00 crc kubenswrapper[4787]: I1001 10:00:00.134584 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321880-mxzcx" Oct 01 10:00:00 crc kubenswrapper[4787]: I1001 10:00:00.136900 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 01 10:00:00 crc kubenswrapper[4787]: I1001 10:00:00.138744 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 01 10:00:00 crc kubenswrapper[4787]: I1001 10:00:00.147316 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321880-mxzcx"] Oct 01 10:00:00 crc kubenswrapper[4787]: I1001 10:00:00.242893 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/21868214-109b-4037-bacc-d984deef93fb-secret-volume\") pod \"collect-profiles-29321880-mxzcx\" (UID: \"21868214-109b-4037-bacc-d984deef93fb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321880-mxzcx" Oct 01 10:00:00 crc kubenswrapper[4787]: I1001 10:00:00.242944 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/21868214-109b-4037-bacc-d984deef93fb-config-volume\") pod \"collect-profiles-29321880-mxzcx\" (UID: \"21868214-109b-4037-bacc-d984deef93fb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321880-mxzcx" Oct 01 10:00:00 crc kubenswrapper[4787]: I1001 10:00:00.243021 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mbwh\" (UniqueName: \"kubernetes.io/projected/21868214-109b-4037-bacc-d984deef93fb-kube-api-access-4mbwh\") pod \"collect-profiles-29321880-mxzcx\" (UID: \"21868214-109b-4037-bacc-d984deef93fb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321880-mxzcx" Oct 01 10:00:00 crc kubenswrapper[4787]: I1001 10:00:00.344146 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/21868214-109b-4037-bacc-d984deef93fb-secret-volume\") pod \"collect-profiles-29321880-mxzcx\" (UID: \"21868214-109b-4037-bacc-d984deef93fb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321880-mxzcx" Oct 01 10:00:00 crc kubenswrapper[4787]: I1001 10:00:00.344206 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/21868214-109b-4037-bacc-d984deef93fb-config-volume\") pod \"collect-profiles-29321880-mxzcx\" (UID: \"21868214-109b-4037-bacc-d984deef93fb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321880-mxzcx" Oct 01 10:00:00 crc kubenswrapper[4787]: I1001 10:00:00.344275 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mbwh\" (UniqueName: \"kubernetes.io/projected/21868214-109b-4037-bacc-d984deef93fb-kube-api-access-4mbwh\") pod \"collect-profiles-29321880-mxzcx\" (UID: \"21868214-109b-4037-bacc-d984deef93fb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321880-mxzcx" Oct 01 10:00:00 crc kubenswrapper[4787]: I1001 10:00:00.346275 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/21868214-109b-4037-bacc-d984deef93fb-config-volume\") pod \"collect-profiles-29321880-mxzcx\" (UID: \"21868214-109b-4037-bacc-d984deef93fb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321880-mxzcx" Oct 01 10:00:00 crc kubenswrapper[4787]: I1001 10:00:00.356513 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/21868214-109b-4037-bacc-d984deef93fb-secret-volume\") pod \"collect-profiles-29321880-mxzcx\" (UID: \"21868214-109b-4037-bacc-d984deef93fb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321880-mxzcx" Oct 01 10:00:00 crc kubenswrapper[4787]: I1001 10:00:00.361651 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mbwh\" (UniqueName: \"kubernetes.io/projected/21868214-109b-4037-bacc-d984deef93fb-kube-api-access-4mbwh\") pod \"collect-profiles-29321880-mxzcx\" (UID: \"21868214-109b-4037-bacc-d984deef93fb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321880-mxzcx" Oct 01 10:00:00 crc kubenswrapper[4787]: I1001 10:00:00.499918 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321880-mxzcx" Oct 01 10:00:00 crc kubenswrapper[4787]: I1001 10:00:00.938554 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321880-mxzcx"] Oct 01 10:00:00 crc kubenswrapper[4787]: W1001 10:00:00.950390 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod21868214_109b_4037_bacc_d984deef93fb.slice/crio-e1f1167061c2d64ab173d3fbd13a52a33f11ad23f65db76fbca880c8cb1e92ad WatchSource:0}: Error finding container e1f1167061c2d64ab173d3fbd13a52a33f11ad23f65db76fbca880c8cb1e92ad: Status 404 returned error can't find the container with id e1f1167061c2d64ab173d3fbd13a52a33f11ad23f65db76fbca880c8cb1e92ad Oct 01 10:00:01 crc kubenswrapper[4787]: I1001 10:00:01.204272 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321880-mxzcx" event={"ID":"21868214-109b-4037-bacc-d984deef93fb","Type":"ContainerStarted","Data":"0bac0d9bdfd0708a883535626a9669c8a625140b488fef50fa60b3c1948cccbf"} Oct 01 10:00:01 crc kubenswrapper[4787]: I1001 10:00:01.204352 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321880-mxzcx" event={"ID":"21868214-109b-4037-bacc-d984deef93fb","Type":"ContainerStarted","Data":"e1f1167061c2d64ab173d3fbd13a52a33f11ad23f65db76fbca880c8cb1e92ad"} Oct 01 10:00:01 crc kubenswrapper[4787]: I1001 10:00:01.225011 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29321880-mxzcx" podStartSLOduration=1.224984186 podStartE2EDuration="1.224984186s" podCreationTimestamp="2025-10-01 10:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 10:00:01.220669491 +0000 UTC m=+1433.335813668" watchObservedRunningTime="2025-10-01 10:00:01.224984186 +0000 UTC m=+1433.340128343" Oct 01 10:00:02 crc kubenswrapper[4787]: I1001 10:00:02.214924 4787 generic.go:334] "Generic (PLEG): container finished" podID="21868214-109b-4037-bacc-d984deef93fb" containerID="0bac0d9bdfd0708a883535626a9669c8a625140b488fef50fa60b3c1948cccbf" exitCode=0 Oct 01 10:00:02 crc kubenswrapper[4787]: I1001 10:00:02.214998 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321880-mxzcx" event={"ID":"21868214-109b-4037-bacc-d984deef93fb","Type":"ContainerDied","Data":"0bac0d9bdfd0708a883535626a9669c8a625140b488fef50fa60b3c1948cccbf"} Oct 01 10:00:03 crc kubenswrapper[4787]: I1001 10:00:03.570798 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321880-mxzcx" Oct 01 10:00:03 crc kubenswrapper[4787]: I1001 10:00:03.618431 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/21868214-109b-4037-bacc-d984deef93fb-secret-volume\") pod \"21868214-109b-4037-bacc-d984deef93fb\" (UID: \"21868214-109b-4037-bacc-d984deef93fb\") " Oct 01 10:00:03 crc kubenswrapper[4787]: I1001 10:00:03.629459 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21868214-109b-4037-bacc-d984deef93fb-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "21868214-109b-4037-bacc-d984deef93fb" (UID: "21868214-109b-4037-bacc-d984deef93fb"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:00:03 crc kubenswrapper[4787]: I1001 10:00:03.720862 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mbwh\" (UniqueName: \"kubernetes.io/projected/21868214-109b-4037-bacc-d984deef93fb-kube-api-access-4mbwh\") pod \"21868214-109b-4037-bacc-d984deef93fb\" (UID: \"21868214-109b-4037-bacc-d984deef93fb\") " Oct 01 10:00:03 crc kubenswrapper[4787]: I1001 10:00:03.721019 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/21868214-109b-4037-bacc-d984deef93fb-config-volume\") pod \"21868214-109b-4037-bacc-d984deef93fb\" (UID: \"21868214-109b-4037-bacc-d984deef93fb\") " Oct 01 10:00:03 crc kubenswrapper[4787]: I1001 10:00:03.721532 4787 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/21868214-109b-4037-bacc-d984deef93fb-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 01 10:00:03 crc kubenswrapper[4787]: I1001 10:00:03.722181 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21868214-109b-4037-bacc-d984deef93fb-config-volume" (OuterVolumeSpecName: "config-volume") pod "21868214-109b-4037-bacc-d984deef93fb" (UID: "21868214-109b-4037-bacc-d984deef93fb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:00:03 crc kubenswrapper[4787]: I1001 10:00:03.723861 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21868214-109b-4037-bacc-d984deef93fb-kube-api-access-4mbwh" (OuterVolumeSpecName: "kube-api-access-4mbwh") pod "21868214-109b-4037-bacc-d984deef93fb" (UID: "21868214-109b-4037-bacc-d984deef93fb"). InnerVolumeSpecName "kube-api-access-4mbwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:00:03 crc kubenswrapper[4787]: I1001 10:00:03.823238 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4mbwh\" (UniqueName: \"kubernetes.io/projected/21868214-109b-4037-bacc-d984deef93fb-kube-api-access-4mbwh\") on node \"crc\" DevicePath \"\"" Oct 01 10:00:03 crc kubenswrapper[4787]: I1001 10:00:03.823273 4787 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/21868214-109b-4037-bacc-d984deef93fb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 01 10:00:04 crc kubenswrapper[4787]: I1001 10:00:04.234944 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321880-mxzcx" event={"ID":"21868214-109b-4037-bacc-d984deef93fb","Type":"ContainerDied","Data":"e1f1167061c2d64ab173d3fbd13a52a33f11ad23f65db76fbca880c8cb1e92ad"} Oct 01 10:00:04 crc kubenswrapper[4787]: I1001 10:00:04.235301 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1f1167061c2d64ab173d3fbd13a52a33f11ad23f65db76fbca880c8cb1e92ad" Oct 01 10:00:04 crc kubenswrapper[4787]: I1001 10:00:04.234996 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321880-mxzcx" Oct 01 10:00:11 crc kubenswrapper[4787]: I1001 10:00:11.250555 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:00:11 crc kubenswrapper[4787]: I1001 10:00:11.251317 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:00:30 crc kubenswrapper[4787]: I1001 10:00:30.410285 4787 scope.go:117] "RemoveContainer" containerID="50854a5e8554c38855622f13e3e07ceedacfd52a6adccf1ebbe564a1a31d2099" Oct 01 10:00:30 crc kubenswrapper[4787]: I1001 10:00:30.444189 4787 scope.go:117] "RemoveContainer" containerID="cd830538ea3ad81ccaddbe633be05f3709972cac61e3d9ad12bb55b9a13bf772" Oct 01 10:00:30 crc kubenswrapper[4787]: I1001 10:00:30.470039 4787 scope.go:117] "RemoveContainer" containerID="2e0b5dd5073c0fa32a8d6df15f44078a84e399426c318ffd3e06018034b44a89" Oct 01 10:00:30 crc kubenswrapper[4787]: I1001 10:00:30.542344 4787 scope.go:117] "RemoveContainer" containerID="fbcdcd11b746a170ee3eb44ccf26a77e4bc1c3687c06637d776d97b81b8f455b" Oct 01 10:00:31 crc kubenswrapper[4787]: I1001 10:00:31.430253 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hxj6h"] Oct 01 10:00:31 crc kubenswrapper[4787]: E1001 10:00:31.430892 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21868214-109b-4037-bacc-d984deef93fb" containerName="collect-profiles" Oct 01 10:00:31 crc kubenswrapper[4787]: I1001 10:00:31.430912 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="21868214-109b-4037-bacc-d984deef93fb" containerName="collect-profiles" Oct 01 10:00:31 crc kubenswrapper[4787]: I1001 10:00:31.431190 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="21868214-109b-4037-bacc-d984deef93fb" containerName="collect-profiles" Oct 01 10:00:31 crc kubenswrapper[4787]: I1001 10:00:31.433360 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hxj6h" Oct 01 10:00:31 crc kubenswrapper[4787]: I1001 10:00:31.441253 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hxj6h"] Oct 01 10:00:31 crc kubenswrapper[4787]: I1001 10:00:31.485711 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5z9sq\" (UniqueName: \"kubernetes.io/projected/c87bc8f7-226f-463b-b49b-06d2efb2a805-kube-api-access-5z9sq\") pod \"redhat-marketplace-hxj6h\" (UID: \"c87bc8f7-226f-463b-b49b-06d2efb2a805\") " pod="openshift-marketplace/redhat-marketplace-hxj6h" Oct 01 10:00:31 crc kubenswrapper[4787]: I1001 10:00:31.485756 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c87bc8f7-226f-463b-b49b-06d2efb2a805-utilities\") pod \"redhat-marketplace-hxj6h\" (UID: \"c87bc8f7-226f-463b-b49b-06d2efb2a805\") " pod="openshift-marketplace/redhat-marketplace-hxj6h" Oct 01 10:00:31 crc kubenswrapper[4787]: I1001 10:00:31.485986 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c87bc8f7-226f-463b-b49b-06d2efb2a805-catalog-content\") pod \"redhat-marketplace-hxj6h\" (UID: \"c87bc8f7-226f-463b-b49b-06d2efb2a805\") " pod="openshift-marketplace/redhat-marketplace-hxj6h" Oct 01 10:00:31 crc kubenswrapper[4787]: I1001 10:00:31.587488 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c87bc8f7-226f-463b-b49b-06d2efb2a805-catalog-content\") pod \"redhat-marketplace-hxj6h\" (UID: \"c87bc8f7-226f-463b-b49b-06d2efb2a805\") " pod="openshift-marketplace/redhat-marketplace-hxj6h" Oct 01 10:00:31 crc kubenswrapper[4787]: I1001 10:00:31.587569 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5z9sq\" (UniqueName: \"kubernetes.io/projected/c87bc8f7-226f-463b-b49b-06d2efb2a805-kube-api-access-5z9sq\") pod \"redhat-marketplace-hxj6h\" (UID: \"c87bc8f7-226f-463b-b49b-06d2efb2a805\") " pod="openshift-marketplace/redhat-marketplace-hxj6h" Oct 01 10:00:31 crc kubenswrapper[4787]: I1001 10:00:31.587588 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c87bc8f7-226f-463b-b49b-06d2efb2a805-utilities\") pod \"redhat-marketplace-hxj6h\" (UID: \"c87bc8f7-226f-463b-b49b-06d2efb2a805\") " pod="openshift-marketplace/redhat-marketplace-hxj6h" Oct 01 10:00:31 crc kubenswrapper[4787]: I1001 10:00:31.590678 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c87bc8f7-226f-463b-b49b-06d2efb2a805-catalog-content\") pod \"redhat-marketplace-hxj6h\" (UID: \"c87bc8f7-226f-463b-b49b-06d2efb2a805\") " pod="openshift-marketplace/redhat-marketplace-hxj6h" Oct 01 10:00:31 crc kubenswrapper[4787]: I1001 10:00:31.590793 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c87bc8f7-226f-463b-b49b-06d2efb2a805-utilities\") pod \"redhat-marketplace-hxj6h\" (UID: \"c87bc8f7-226f-463b-b49b-06d2efb2a805\") " pod="openshift-marketplace/redhat-marketplace-hxj6h" Oct 01 10:00:31 crc kubenswrapper[4787]: I1001 10:00:31.608879 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5z9sq\" (UniqueName: \"kubernetes.io/projected/c87bc8f7-226f-463b-b49b-06d2efb2a805-kube-api-access-5z9sq\") pod \"redhat-marketplace-hxj6h\" (UID: \"c87bc8f7-226f-463b-b49b-06d2efb2a805\") " pod="openshift-marketplace/redhat-marketplace-hxj6h" Oct 01 10:00:31 crc kubenswrapper[4787]: I1001 10:00:31.762940 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hxj6h" Oct 01 10:00:32 crc kubenswrapper[4787]: I1001 10:00:32.243859 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hxj6h"] Oct 01 10:00:32 crc kubenswrapper[4787]: I1001 10:00:32.543618 4787 generic.go:334] "Generic (PLEG): container finished" podID="c87bc8f7-226f-463b-b49b-06d2efb2a805" containerID="2a6217c9df0e838589e52974f5f7638322951f961cad9bf22b0653eaf81070fd" exitCode=0 Oct 01 10:00:32 crc kubenswrapper[4787]: I1001 10:00:32.543686 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hxj6h" event={"ID":"c87bc8f7-226f-463b-b49b-06d2efb2a805","Type":"ContainerDied","Data":"2a6217c9df0e838589e52974f5f7638322951f961cad9bf22b0653eaf81070fd"} Oct 01 10:00:32 crc kubenswrapper[4787]: I1001 10:00:32.544193 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hxj6h" event={"ID":"c87bc8f7-226f-463b-b49b-06d2efb2a805","Type":"ContainerStarted","Data":"66a9346832e013fbd9c468d9a55f4e58ccc36420f414366d8a9395a72518eea4"} Oct 01 10:00:33 crc kubenswrapper[4787]: I1001 10:00:33.554535 4787 generic.go:334] "Generic (PLEG): container finished" podID="c87bc8f7-226f-463b-b49b-06d2efb2a805" containerID="65d74f23d64972c7a13da501085b4058a711d1f11f29f89506ad5646df36ae79" exitCode=0 Oct 01 10:00:33 crc kubenswrapper[4787]: I1001 10:00:33.554583 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hxj6h" event={"ID":"c87bc8f7-226f-463b-b49b-06d2efb2a805","Type":"ContainerDied","Data":"65d74f23d64972c7a13da501085b4058a711d1f11f29f89506ad5646df36ae79"} Oct 01 10:00:34 crc kubenswrapper[4787]: I1001 10:00:34.567237 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hxj6h" event={"ID":"c87bc8f7-226f-463b-b49b-06d2efb2a805","Type":"ContainerStarted","Data":"93db8df24ff6af6ff517250b0923df66aaf9351933c68994db0135609be91d21"} Oct 01 10:00:34 crc kubenswrapper[4787]: I1001 10:00:34.588756 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hxj6h" podStartSLOduration=2.159718579 podStartE2EDuration="3.588736154s" podCreationTimestamp="2025-10-01 10:00:31 +0000 UTC" firstStartedPulling="2025-10-01 10:00:32.545568002 +0000 UTC m=+1464.660712159" lastFinishedPulling="2025-10-01 10:00:33.974585577 +0000 UTC m=+1466.089729734" observedRunningTime="2025-10-01 10:00:34.588015326 +0000 UTC m=+1466.703159493" watchObservedRunningTime="2025-10-01 10:00:34.588736154 +0000 UTC m=+1466.703880311" Oct 01 10:00:41 crc kubenswrapper[4787]: I1001 10:00:41.250536 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:00:41 crc kubenswrapper[4787]: I1001 10:00:41.251307 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:00:41 crc kubenswrapper[4787]: I1001 10:00:41.251409 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" Oct 01 10:00:41 crc kubenswrapper[4787]: I1001 10:00:41.252428 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e01333312cf21546b537b606d139d2d34e67fa8885c08f9ceb1f4c790ae4fd3a"} pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 10:00:41 crc kubenswrapper[4787]: I1001 10:00:41.252525 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" containerID="cri-o://e01333312cf21546b537b606d139d2d34e67fa8885c08f9ceb1f4c790ae4fd3a" gracePeriod=600 Oct 01 10:00:41 crc kubenswrapper[4787]: E1001 10:00:41.374468 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:00:41 crc kubenswrapper[4787]: I1001 10:00:41.651620 4787 generic.go:334] "Generic (PLEG): container finished" podID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerID="e01333312cf21546b537b606d139d2d34e67fa8885c08f9ceb1f4c790ae4fd3a" exitCode=0 Oct 01 10:00:41 crc kubenswrapper[4787]: I1001 10:00:41.651665 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" event={"ID":"fa51b883-5f3e-4141-9d4c-37704aac7718","Type":"ContainerDied","Data":"e01333312cf21546b537b606d139d2d34e67fa8885c08f9ceb1f4c790ae4fd3a"} Oct 01 10:00:41 crc kubenswrapper[4787]: I1001 10:00:41.651697 4787 scope.go:117] "RemoveContainer" containerID="a85043035562ba4be7fcc084d6a9297752334c5b023237849bd5d1037f591d75" Oct 01 10:00:41 crc kubenswrapper[4787]: I1001 10:00:41.652402 4787 scope.go:117] "RemoveContainer" containerID="e01333312cf21546b537b606d139d2d34e67fa8885c08f9ceb1f4c790ae4fd3a" Oct 01 10:00:41 crc kubenswrapper[4787]: E1001 10:00:41.652678 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:00:41 crc kubenswrapper[4787]: I1001 10:00:41.763889 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hxj6h" Oct 01 10:00:41 crc kubenswrapper[4787]: I1001 10:00:41.764248 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hxj6h" Oct 01 10:00:41 crc kubenswrapper[4787]: I1001 10:00:41.813016 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hxj6h" Oct 01 10:00:42 crc kubenswrapper[4787]: I1001 10:00:42.720788 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hxj6h" Oct 01 10:00:42 crc kubenswrapper[4787]: I1001 10:00:42.795435 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hxj6h"] Oct 01 10:00:44 crc kubenswrapper[4787]: I1001 10:00:44.684867 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hxj6h" podUID="c87bc8f7-226f-463b-b49b-06d2efb2a805" containerName="registry-server" containerID="cri-o://93db8df24ff6af6ff517250b0923df66aaf9351933c68994db0135609be91d21" gracePeriod=2 Oct 01 10:00:45 crc kubenswrapper[4787]: I1001 10:00:45.210714 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hxj6h" Oct 01 10:00:45 crc kubenswrapper[4787]: I1001 10:00:45.266433 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c87bc8f7-226f-463b-b49b-06d2efb2a805-utilities\") pod \"c87bc8f7-226f-463b-b49b-06d2efb2a805\" (UID: \"c87bc8f7-226f-463b-b49b-06d2efb2a805\") " Oct 01 10:00:45 crc kubenswrapper[4787]: I1001 10:00:45.266672 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c87bc8f7-226f-463b-b49b-06d2efb2a805-catalog-content\") pod \"c87bc8f7-226f-463b-b49b-06d2efb2a805\" (UID: \"c87bc8f7-226f-463b-b49b-06d2efb2a805\") " Oct 01 10:00:45 crc kubenswrapper[4787]: I1001 10:00:45.266771 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5z9sq\" (UniqueName: \"kubernetes.io/projected/c87bc8f7-226f-463b-b49b-06d2efb2a805-kube-api-access-5z9sq\") pod \"c87bc8f7-226f-463b-b49b-06d2efb2a805\" (UID: \"c87bc8f7-226f-463b-b49b-06d2efb2a805\") " Oct 01 10:00:45 crc kubenswrapper[4787]: I1001 10:00:45.269018 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c87bc8f7-226f-463b-b49b-06d2efb2a805-utilities" (OuterVolumeSpecName: "utilities") pod "c87bc8f7-226f-463b-b49b-06d2efb2a805" (UID: "c87bc8f7-226f-463b-b49b-06d2efb2a805"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:00:45 crc kubenswrapper[4787]: I1001 10:00:45.276505 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c87bc8f7-226f-463b-b49b-06d2efb2a805-kube-api-access-5z9sq" (OuterVolumeSpecName: "kube-api-access-5z9sq") pod "c87bc8f7-226f-463b-b49b-06d2efb2a805" (UID: "c87bc8f7-226f-463b-b49b-06d2efb2a805"). InnerVolumeSpecName "kube-api-access-5z9sq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:00:45 crc kubenswrapper[4787]: I1001 10:00:45.281304 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c87bc8f7-226f-463b-b49b-06d2efb2a805-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c87bc8f7-226f-463b-b49b-06d2efb2a805" (UID: "c87bc8f7-226f-463b-b49b-06d2efb2a805"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:00:45 crc kubenswrapper[4787]: I1001 10:00:45.369835 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c87bc8f7-226f-463b-b49b-06d2efb2a805-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 10:00:45 crc kubenswrapper[4787]: I1001 10:00:45.369872 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5z9sq\" (UniqueName: \"kubernetes.io/projected/c87bc8f7-226f-463b-b49b-06d2efb2a805-kube-api-access-5z9sq\") on node \"crc\" DevicePath \"\"" Oct 01 10:00:45 crc kubenswrapper[4787]: I1001 10:00:45.369884 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c87bc8f7-226f-463b-b49b-06d2efb2a805-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 10:00:45 crc kubenswrapper[4787]: I1001 10:00:45.697411 4787 generic.go:334] "Generic (PLEG): container finished" podID="c87bc8f7-226f-463b-b49b-06d2efb2a805" containerID="93db8df24ff6af6ff517250b0923df66aaf9351933c68994db0135609be91d21" exitCode=0 Oct 01 10:00:45 crc kubenswrapper[4787]: I1001 10:00:45.697460 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hxj6h" event={"ID":"c87bc8f7-226f-463b-b49b-06d2efb2a805","Type":"ContainerDied","Data":"93db8df24ff6af6ff517250b0923df66aaf9351933c68994db0135609be91d21"} Oct 01 10:00:45 crc kubenswrapper[4787]: I1001 10:00:45.697523 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hxj6h" Oct 01 10:00:45 crc kubenswrapper[4787]: I1001 10:00:45.697553 4787 scope.go:117] "RemoveContainer" containerID="93db8df24ff6af6ff517250b0923df66aaf9351933c68994db0135609be91d21" Oct 01 10:00:45 crc kubenswrapper[4787]: I1001 10:00:45.697539 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hxj6h" event={"ID":"c87bc8f7-226f-463b-b49b-06d2efb2a805","Type":"ContainerDied","Data":"66a9346832e013fbd9c468d9a55f4e58ccc36420f414366d8a9395a72518eea4"} Oct 01 10:00:45 crc kubenswrapper[4787]: I1001 10:00:45.736974 4787 scope.go:117] "RemoveContainer" containerID="65d74f23d64972c7a13da501085b4058a711d1f11f29f89506ad5646df36ae79" Oct 01 10:00:45 crc kubenswrapper[4787]: I1001 10:00:45.742744 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hxj6h"] Oct 01 10:00:45 crc kubenswrapper[4787]: I1001 10:00:45.754540 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hxj6h"] Oct 01 10:00:45 crc kubenswrapper[4787]: I1001 10:00:45.766327 4787 scope.go:117] "RemoveContainer" containerID="2a6217c9df0e838589e52974f5f7638322951f961cad9bf22b0653eaf81070fd" Oct 01 10:00:45 crc kubenswrapper[4787]: I1001 10:00:45.826554 4787 scope.go:117] "RemoveContainer" containerID="93db8df24ff6af6ff517250b0923df66aaf9351933c68994db0135609be91d21" Oct 01 10:00:45 crc kubenswrapper[4787]: E1001 10:00:45.827473 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93db8df24ff6af6ff517250b0923df66aaf9351933c68994db0135609be91d21\": container with ID starting with 93db8df24ff6af6ff517250b0923df66aaf9351933c68994db0135609be91d21 not found: ID does not exist" containerID="93db8df24ff6af6ff517250b0923df66aaf9351933c68994db0135609be91d21" Oct 01 10:00:45 crc kubenswrapper[4787]: I1001 10:00:45.827603 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93db8df24ff6af6ff517250b0923df66aaf9351933c68994db0135609be91d21"} err="failed to get container status \"93db8df24ff6af6ff517250b0923df66aaf9351933c68994db0135609be91d21\": rpc error: code = NotFound desc = could not find container \"93db8df24ff6af6ff517250b0923df66aaf9351933c68994db0135609be91d21\": container with ID starting with 93db8df24ff6af6ff517250b0923df66aaf9351933c68994db0135609be91d21 not found: ID does not exist" Oct 01 10:00:45 crc kubenswrapper[4787]: I1001 10:00:45.827696 4787 scope.go:117] "RemoveContainer" containerID="65d74f23d64972c7a13da501085b4058a711d1f11f29f89506ad5646df36ae79" Oct 01 10:00:45 crc kubenswrapper[4787]: E1001 10:00:45.828104 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65d74f23d64972c7a13da501085b4058a711d1f11f29f89506ad5646df36ae79\": container with ID starting with 65d74f23d64972c7a13da501085b4058a711d1f11f29f89506ad5646df36ae79 not found: ID does not exist" containerID="65d74f23d64972c7a13da501085b4058a711d1f11f29f89506ad5646df36ae79" Oct 01 10:00:45 crc kubenswrapper[4787]: I1001 10:00:45.828142 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65d74f23d64972c7a13da501085b4058a711d1f11f29f89506ad5646df36ae79"} err="failed to get container status \"65d74f23d64972c7a13da501085b4058a711d1f11f29f89506ad5646df36ae79\": rpc error: code = NotFound desc = could not find container \"65d74f23d64972c7a13da501085b4058a711d1f11f29f89506ad5646df36ae79\": container with ID starting with 65d74f23d64972c7a13da501085b4058a711d1f11f29f89506ad5646df36ae79 not found: ID does not exist" Oct 01 10:00:45 crc kubenswrapper[4787]: I1001 10:00:45.828171 4787 scope.go:117] "RemoveContainer" containerID="2a6217c9df0e838589e52974f5f7638322951f961cad9bf22b0653eaf81070fd" Oct 01 10:00:45 crc kubenswrapper[4787]: E1001 10:00:45.828402 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a6217c9df0e838589e52974f5f7638322951f961cad9bf22b0653eaf81070fd\": container with ID starting with 2a6217c9df0e838589e52974f5f7638322951f961cad9bf22b0653eaf81070fd not found: ID does not exist" containerID="2a6217c9df0e838589e52974f5f7638322951f961cad9bf22b0653eaf81070fd" Oct 01 10:00:45 crc kubenswrapper[4787]: I1001 10:00:45.828430 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a6217c9df0e838589e52974f5f7638322951f961cad9bf22b0653eaf81070fd"} err="failed to get container status \"2a6217c9df0e838589e52974f5f7638322951f961cad9bf22b0653eaf81070fd\": rpc error: code = NotFound desc = could not find container \"2a6217c9df0e838589e52974f5f7638322951f961cad9bf22b0653eaf81070fd\": container with ID starting with 2a6217c9df0e838589e52974f5f7638322951f961cad9bf22b0653eaf81070fd not found: ID does not exist" Oct 01 10:00:46 crc kubenswrapper[4787]: I1001 10:00:46.546858 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c87bc8f7-226f-463b-b49b-06d2efb2a805" path="/var/lib/kubelet/pods/c87bc8f7-226f-463b-b49b-06d2efb2a805/volumes" Oct 01 10:00:54 crc kubenswrapper[4787]: I1001 10:00:54.524674 4787 scope.go:117] "RemoveContainer" containerID="e01333312cf21546b537b606d139d2d34e67fa8885c08f9ceb1f4c790ae4fd3a" Oct 01 10:00:54 crc kubenswrapper[4787]: E1001 10:00:54.526020 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:01:00 crc kubenswrapper[4787]: I1001 10:01:00.148054 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29321881-dnxcs"] Oct 01 10:01:00 crc kubenswrapper[4787]: E1001 10:01:00.149023 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c87bc8f7-226f-463b-b49b-06d2efb2a805" containerName="registry-server" Oct 01 10:01:00 crc kubenswrapper[4787]: I1001 10:01:00.149038 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c87bc8f7-226f-463b-b49b-06d2efb2a805" containerName="registry-server" Oct 01 10:01:00 crc kubenswrapper[4787]: E1001 10:01:00.149057 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c87bc8f7-226f-463b-b49b-06d2efb2a805" containerName="extract-content" Oct 01 10:01:00 crc kubenswrapper[4787]: I1001 10:01:00.149064 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c87bc8f7-226f-463b-b49b-06d2efb2a805" containerName="extract-content" Oct 01 10:01:00 crc kubenswrapper[4787]: E1001 10:01:00.149109 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c87bc8f7-226f-463b-b49b-06d2efb2a805" containerName="extract-utilities" Oct 01 10:01:00 crc kubenswrapper[4787]: I1001 10:01:00.149121 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c87bc8f7-226f-463b-b49b-06d2efb2a805" containerName="extract-utilities" Oct 01 10:01:00 crc kubenswrapper[4787]: I1001 10:01:00.149334 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="c87bc8f7-226f-463b-b49b-06d2efb2a805" containerName="registry-server" Oct 01 10:01:00 crc kubenswrapper[4787]: I1001 10:01:00.150066 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29321881-dnxcs" Oct 01 10:01:00 crc kubenswrapper[4787]: I1001 10:01:00.158827 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29321881-dnxcs"] Oct 01 10:01:00 crc kubenswrapper[4787]: I1001 10:01:00.229479 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7b0f7d9e-ab06-41c1-bac1-54004f737044-fernet-keys\") pod \"keystone-cron-29321881-dnxcs\" (UID: \"7b0f7d9e-ab06-41c1-bac1-54004f737044\") " pod="openstack/keystone-cron-29321881-dnxcs" Oct 01 10:01:00 crc kubenswrapper[4787]: I1001 10:01:00.229795 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ns225\" (UniqueName: \"kubernetes.io/projected/7b0f7d9e-ab06-41c1-bac1-54004f737044-kube-api-access-ns225\") pod \"keystone-cron-29321881-dnxcs\" (UID: \"7b0f7d9e-ab06-41c1-bac1-54004f737044\") " pod="openstack/keystone-cron-29321881-dnxcs" Oct 01 10:01:00 crc kubenswrapper[4787]: I1001 10:01:00.229903 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b0f7d9e-ab06-41c1-bac1-54004f737044-config-data\") pod \"keystone-cron-29321881-dnxcs\" (UID: \"7b0f7d9e-ab06-41c1-bac1-54004f737044\") " pod="openstack/keystone-cron-29321881-dnxcs" Oct 01 10:01:00 crc kubenswrapper[4787]: I1001 10:01:00.229987 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b0f7d9e-ab06-41c1-bac1-54004f737044-combined-ca-bundle\") pod \"keystone-cron-29321881-dnxcs\" (UID: \"7b0f7d9e-ab06-41c1-bac1-54004f737044\") " pod="openstack/keystone-cron-29321881-dnxcs" Oct 01 10:01:00 crc kubenswrapper[4787]: I1001 10:01:00.332492 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ns225\" (UniqueName: \"kubernetes.io/projected/7b0f7d9e-ab06-41c1-bac1-54004f737044-kube-api-access-ns225\") pod \"keystone-cron-29321881-dnxcs\" (UID: \"7b0f7d9e-ab06-41c1-bac1-54004f737044\") " pod="openstack/keystone-cron-29321881-dnxcs" Oct 01 10:01:00 crc kubenswrapper[4787]: I1001 10:01:00.332616 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b0f7d9e-ab06-41c1-bac1-54004f737044-config-data\") pod \"keystone-cron-29321881-dnxcs\" (UID: \"7b0f7d9e-ab06-41c1-bac1-54004f737044\") " pod="openstack/keystone-cron-29321881-dnxcs" Oct 01 10:01:00 crc kubenswrapper[4787]: I1001 10:01:00.332650 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b0f7d9e-ab06-41c1-bac1-54004f737044-combined-ca-bundle\") pod \"keystone-cron-29321881-dnxcs\" (UID: \"7b0f7d9e-ab06-41c1-bac1-54004f737044\") " pod="openstack/keystone-cron-29321881-dnxcs" Oct 01 10:01:00 crc kubenswrapper[4787]: I1001 10:01:00.332812 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7b0f7d9e-ab06-41c1-bac1-54004f737044-fernet-keys\") pod \"keystone-cron-29321881-dnxcs\" (UID: \"7b0f7d9e-ab06-41c1-bac1-54004f737044\") " pod="openstack/keystone-cron-29321881-dnxcs" Oct 01 10:01:00 crc kubenswrapper[4787]: I1001 10:01:00.339802 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7b0f7d9e-ab06-41c1-bac1-54004f737044-fernet-keys\") pod \"keystone-cron-29321881-dnxcs\" (UID: \"7b0f7d9e-ab06-41c1-bac1-54004f737044\") " pod="openstack/keystone-cron-29321881-dnxcs" Oct 01 10:01:00 crc kubenswrapper[4787]: I1001 10:01:00.340173 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b0f7d9e-ab06-41c1-bac1-54004f737044-config-data\") pod \"keystone-cron-29321881-dnxcs\" (UID: \"7b0f7d9e-ab06-41c1-bac1-54004f737044\") " pod="openstack/keystone-cron-29321881-dnxcs" Oct 01 10:01:00 crc kubenswrapper[4787]: I1001 10:01:00.341387 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b0f7d9e-ab06-41c1-bac1-54004f737044-combined-ca-bundle\") pod \"keystone-cron-29321881-dnxcs\" (UID: \"7b0f7d9e-ab06-41c1-bac1-54004f737044\") " pod="openstack/keystone-cron-29321881-dnxcs" Oct 01 10:01:00 crc kubenswrapper[4787]: I1001 10:01:00.360336 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ns225\" (UniqueName: \"kubernetes.io/projected/7b0f7d9e-ab06-41c1-bac1-54004f737044-kube-api-access-ns225\") pod \"keystone-cron-29321881-dnxcs\" (UID: \"7b0f7d9e-ab06-41c1-bac1-54004f737044\") " pod="openstack/keystone-cron-29321881-dnxcs" Oct 01 10:01:00 crc kubenswrapper[4787]: I1001 10:01:00.477811 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29321881-dnxcs" Oct 01 10:01:00 crc kubenswrapper[4787]: I1001 10:01:00.938670 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29321881-dnxcs"] Oct 01 10:01:01 crc kubenswrapper[4787]: I1001 10:01:01.887489 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29321881-dnxcs" event={"ID":"7b0f7d9e-ab06-41c1-bac1-54004f737044","Type":"ContainerStarted","Data":"cfd2a9f271ff392b3d59286f1533539c623f8e4f636e4d3110ad51acc2d0be5c"} Oct 01 10:01:01 crc kubenswrapper[4787]: I1001 10:01:01.887983 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29321881-dnxcs" event={"ID":"7b0f7d9e-ab06-41c1-bac1-54004f737044","Type":"ContainerStarted","Data":"fae5e6c9f912d733f4616594e63b6d85bafb513585c780aca3ff93c5246a19f1"} Oct 01 10:01:01 crc kubenswrapper[4787]: I1001 10:01:01.923357 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29321881-dnxcs" podStartSLOduration=1.923327335 podStartE2EDuration="1.923327335s" podCreationTimestamp="2025-10-01 10:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 10:01:01.915053845 +0000 UTC m=+1494.030198032" watchObservedRunningTime="2025-10-01 10:01:01.923327335 +0000 UTC m=+1494.038471512" Oct 01 10:01:03 crc kubenswrapper[4787]: I1001 10:01:03.906972 4787 generic.go:334] "Generic (PLEG): container finished" podID="7b0f7d9e-ab06-41c1-bac1-54004f737044" containerID="cfd2a9f271ff392b3d59286f1533539c623f8e4f636e4d3110ad51acc2d0be5c" exitCode=0 Oct 01 10:01:03 crc kubenswrapper[4787]: I1001 10:01:03.907070 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29321881-dnxcs" event={"ID":"7b0f7d9e-ab06-41c1-bac1-54004f737044","Type":"ContainerDied","Data":"cfd2a9f271ff392b3d59286f1533539c623f8e4f636e4d3110ad51acc2d0be5c"} Oct 01 10:01:05 crc kubenswrapper[4787]: I1001 10:01:05.303991 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29321881-dnxcs" Oct 01 10:01:05 crc kubenswrapper[4787]: I1001 10:01:05.457709 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ns225\" (UniqueName: \"kubernetes.io/projected/7b0f7d9e-ab06-41c1-bac1-54004f737044-kube-api-access-ns225\") pod \"7b0f7d9e-ab06-41c1-bac1-54004f737044\" (UID: \"7b0f7d9e-ab06-41c1-bac1-54004f737044\") " Oct 01 10:01:05 crc kubenswrapper[4787]: I1001 10:01:05.457777 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b0f7d9e-ab06-41c1-bac1-54004f737044-config-data\") pod \"7b0f7d9e-ab06-41c1-bac1-54004f737044\" (UID: \"7b0f7d9e-ab06-41c1-bac1-54004f737044\") " Oct 01 10:01:05 crc kubenswrapper[4787]: I1001 10:01:05.457879 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7b0f7d9e-ab06-41c1-bac1-54004f737044-fernet-keys\") pod \"7b0f7d9e-ab06-41c1-bac1-54004f737044\" (UID: \"7b0f7d9e-ab06-41c1-bac1-54004f737044\") " Oct 01 10:01:05 crc kubenswrapper[4787]: I1001 10:01:05.457939 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b0f7d9e-ab06-41c1-bac1-54004f737044-combined-ca-bundle\") pod \"7b0f7d9e-ab06-41c1-bac1-54004f737044\" (UID: \"7b0f7d9e-ab06-41c1-bac1-54004f737044\") " Oct 01 10:01:05 crc kubenswrapper[4787]: I1001 10:01:05.464255 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b0f7d9e-ab06-41c1-bac1-54004f737044-kube-api-access-ns225" (OuterVolumeSpecName: "kube-api-access-ns225") pod "7b0f7d9e-ab06-41c1-bac1-54004f737044" (UID: "7b0f7d9e-ab06-41c1-bac1-54004f737044"). InnerVolumeSpecName "kube-api-access-ns225". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:01:05 crc kubenswrapper[4787]: I1001 10:01:05.465202 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b0f7d9e-ab06-41c1-bac1-54004f737044-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "7b0f7d9e-ab06-41c1-bac1-54004f737044" (UID: "7b0f7d9e-ab06-41c1-bac1-54004f737044"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:01:05 crc kubenswrapper[4787]: I1001 10:01:05.486568 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b0f7d9e-ab06-41c1-bac1-54004f737044-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7b0f7d9e-ab06-41c1-bac1-54004f737044" (UID: "7b0f7d9e-ab06-41c1-bac1-54004f737044"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:01:05 crc kubenswrapper[4787]: I1001 10:01:05.508020 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b0f7d9e-ab06-41c1-bac1-54004f737044-config-data" (OuterVolumeSpecName: "config-data") pod "7b0f7d9e-ab06-41c1-bac1-54004f737044" (UID: "7b0f7d9e-ab06-41c1-bac1-54004f737044"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:01:05 crc kubenswrapper[4787]: I1001 10:01:05.560532 4787 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7b0f7d9e-ab06-41c1-bac1-54004f737044-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 01 10:01:05 crc kubenswrapper[4787]: I1001 10:01:05.560773 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b0f7d9e-ab06-41c1-bac1-54004f737044-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 10:01:05 crc kubenswrapper[4787]: I1001 10:01:05.560848 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ns225\" (UniqueName: \"kubernetes.io/projected/7b0f7d9e-ab06-41c1-bac1-54004f737044-kube-api-access-ns225\") on node \"crc\" DevicePath \"\"" Oct 01 10:01:05 crc kubenswrapper[4787]: I1001 10:01:05.560926 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b0f7d9e-ab06-41c1-bac1-54004f737044-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 10:01:05 crc kubenswrapper[4787]: I1001 10:01:05.930853 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29321881-dnxcs" event={"ID":"7b0f7d9e-ab06-41c1-bac1-54004f737044","Type":"ContainerDied","Data":"fae5e6c9f912d733f4616594e63b6d85bafb513585c780aca3ff93c5246a19f1"} Oct 01 10:01:05 crc kubenswrapper[4787]: I1001 10:01:05.931288 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fae5e6c9f912d733f4616594e63b6d85bafb513585c780aca3ff93c5246a19f1" Oct 01 10:01:05 crc kubenswrapper[4787]: I1001 10:01:05.930913 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29321881-dnxcs" Oct 01 10:01:06 crc kubenswrapper[4787]: I1001 10:01:06.524328 4787 scope.go:117] "RemoveContainer" containerID="e01333312cf21546b537b606d139d2d34e67fa8885c08f9ceb1f4c790ae4fd3a" Oct 01 10:01:06 crc kubenswrapper[4787]: E1001 10:01:06.524627 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:01:21 crc kubenswrapper[4787]: I1001 10:01:21.523784 4787 scope.go:117] "RemoveContainer" containerID="e01333312cf21546b537b606d139d2d34e67fa8885c08f9ceb1f4c790ae4fd3a" Oct 01 10:01:21 crc kubenswrapper[4787]: E1001 10:01:21.524699 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:01:25 crc kubenswrapper[4787]: I1001 10:01:25.050546 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-4lszz"] Oct 01 10:01:25 crc kubenswrapper[4787]: I1001 10:01:25.081201 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-67phv"] Oct 01 10:01:25 crc kubenswrapper[4787]: I1001 10:01:25.092852 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-67phv"] Oct 01 10:01:25 crc kubenswrapper[4787]: I1001 10:01:25.101553 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-4lszz"] Oct 01 10:01:25 crc kubenswrapper[4787]: I1001 10:01:25.118615 4787 generic.go:334] "Generic (PLEG): container finished" podID="f70ffa29-bf25-44a2-bd45-822e5f59a5d5" containerID="7cf3e48f00ff3dc03265b85bf5517015639767c21f22edd24096e676468b2266" exitCode=0 Oct 01 10:01:25 crc kubenswrapper[4787]: I1001 10:01:25.118686 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4jp68" event={"ID":"f70ffa29-bf25-44a2-bd45-822e5f59a5d5","Type":"ContainerDied","Data":"7cf3e48f00ff3dc03265b85bf5517015639767c21f22edd24096e676468b2266"} Oct 01 10:01:26 crc kubenswrapper[4787]: I1001 10:01:26.538403 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11413a2c-48bf-4160-8e54-d843305b3b97" path="/var/lib/kubelet/pods/11413a2c-48bf-4160-8e54-d843305b3b97/volumes" Oct 01 10:01:26 crc kubenswrapper[4787]: I1001 10:01:26.540146 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29fb6825-d8d1-4401-b0ed-e1f3538cf4c5" path="/var/lib/kubelet/pods/29fb6825-d8d1-4401-b0ed-e1f3538cf4c5/volumes" Oct 01 10:01:26 crc kubenswrapper[4787]: I1001 10:01:26.633352 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4jp68" Oct 01 10:01:26 crc kubenswrapper[4787]: I1001 10:01:26.671867 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f70ffa29-bf25-44a2-bd45-822e5f59a5d5-inventory\") pod \"f70ffa29-bf25-44a2-bd45-822e5f59a5d5\" (UID: \"f70ffa29-bf25-44a2-bd45-822e5f59a5d5\") " Oct 01 10:01:26 crc kubenswrapper[4787]: I1001 10:01:26.672143 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58tln\" (UniqueName: \"kubernetes.io/projected/f70ffa29-bf25-44a2-bd45-822e5f59a5d5-kube-api-access-58tln\") pod \"f70ffa29-bf25-44a2-bd45-822e5f59a5d5\" (UID: \"f70ffa29-bf25-44a2-bd45-822e5f59a5d5\") " Oct 01 10:01:26 crc kubenswrapper[4787]: I1001 10:01:26.672182 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f70ffa29-bf25-44a2-bd45-822e5f59a5d5-ssh-key\") pod \"f70ffa29-bf25-44a2-bd45-822e5f59a5d5\" (UID: \"f70ffa29-bf25-44a2-bd45-822e5f59a5d5\") " Oct 01 10:01:26 crc kubenswrapper[4787]: I1001 10:01:26.678632 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f70ffa29-bf25-44a2-bd45-822e5f59a5d5-kube-api-access-58tln" (OuterVolumeSpecName: "kube-api-access-58tln") pod "f70ffa29-bf25-44a2-bd45-822e5f59a5d5" (UID: "f70ffa29-bf25-44a2-bd45-822e5f59a5d5"). InnerVolumeSpecName "kube-api-access-58tln". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:01:26 crc kubenswrapper[4787]: I1001 10:01:26.701249 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f70ffa29-bf25-44a2-bd45-822e5f59a5d5-inventory" (OuterVolumeSpecName: "inventory") pod "f70ffa29-bf25-44a2-bd45-822e5f59a5d5" (UID: "f70ffa29-bf25-44a2-bd45-822e5f59a5d5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:01:26 crc kubenswrapper[4787]: I1001 10:01:26.701472 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f70ffa29-bf25-44a2-bd45-822e5f59a5d5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f70ffa29-bf25-44a2-bd45-822e5f59a5d5" (UID: "f70ffa29-bf25-44a2-bd45-822e5f59a5d5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:01:26 crc kubenswrapper[4787]: I1001 10:01:26.778238 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f70ffa29-bf25-44a2-bd45-822e5f59a5d5-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 10:01:26 crc kubenswrapper[4787]: I1001 10:01:26.778273 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58tln\" (UniqueName: \"kubernetes.io/projected/f70ffa29-bf25-44a2-bd45-822e5f59a5d5-kube-api-access-58tln\") on node \"crc\" DevicePath \"\"" Oct 01 10:01:26 crc kubenswrapper[4787]: I1001 10:01:26.778288 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f70ffa29-bf25-44a2-bd45-822e5f59a5d5-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 10:01:27 crc kubenswrapper[4787]: I1001 10:01:27.144788 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4jp68" event={"ID":"f70ffa29-bf25-44a2-bd45-822e5f59a5d5","Type":"ContainerDied","Data":"33d40988c9c35849c4c440ced9ca6bbad3802d44bbadfe74266adfc65ec8eda7"} Oct 01 10:01:27 crc kubenswrapper[4787]: I1001 10:01:27.145234 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33d40988c9c35849c4c440ced9ca6bbad3802d44bbadfe74266adfc65ec8eda7" Oct 01 10:01:27 crc kubenswrapper[4787]: I1001 10:01:27.144864 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4jp68" Oct 01 10:01:27 crc kubenswrapper[4787]: I1001 10:01:27.241563 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-llpfd"] Oct 01 10:01:27 crc kubenswrapper[4787]: E1001 10:01:27.242041 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f70ffa29-bf25-44a2-bd45-822e5f59a5d5" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 01 10:01:27 crc kubenswrapper[4787]: I1001 10:01:27.242063 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="f70ffa29-bf25-44a2-bd45-822e5f59a5d5" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 01 10:01:27 crc kubenswrapper[4787]: E1001 10:01:27.242097 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b0f7d9e-ab06-41c1-bac1-54004f737044" containerName="keystone-cron" Oct 01 10:01:27 crc kubenswrapper[4787]: I1001 10:01:27.242107 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b0f7d9e-ab06-41c1-bac1-54004f737044" containerName="keystone-cron" Oct 01 10:01:27 crc kubenswrapper[4787]: I1001 10:01:27.242316 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b0f7d9e-ab06-41c1-bac1-54004f737044" containerName="keystone-cron" Oct 01 10:01:27 crc kubenswrapper[4787]: I1001 10:01:27.242347 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="f70ffa29-bf25-44a2-bd45-822e5f59a5d5" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 01 10:01:27 crc kubenswrapper[4787]: I1001 10:01:27.243031 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-llpfd" Oct 01 10:01:27 crc kubenswrapper[4787]: I1001 10:01:27.246309 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-llpfd"] Oct 01 10:01:27 crc kubenswrapper[4787]: I1001 10:01:27.249502 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 10:01:27 crc kubenswrapper[4787]: I1001 10:01:27.249572 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 10:01:27 crc kubenswrapper[4787]: I1001 10:01:27.249576 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 10:01:27 crc kubenswrapper[4787]: I1001 10:01:27.250021 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vh5v6" Oct 01 10:01:27 crc kubenswrapper[4787]: I1001 10:01:27.388252 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8dd2f969-9904-4848-b0e5-f852a8b958ea-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-llpfd\" (UID: \"8dd2f969-9904-4848-b0e5-f852a8b958ea\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-llpfd" Oct 01 10:01:27 crc kubenswrapper[4787]: I1001 10:01:27.388404 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8dd2f969-9904-4848-b0e5-f852a8b958ea-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-llpfd\" (UID: \"8dd2f969-9904-4848-b0e5-f852a8b958ea\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-llpfd" Oct 01 10:01:27 crc kubenswrapper[4787]: I1001 10:01:27.388558 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sw9z\" (UniqueName: \"kubernetes.io/projected/8dd2f969-9904-4848-b0e5-f852a8b958ea-kube-api-access-5sw9z\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-llpfd\" (UID: \"8dd2f969-9904-4848-b0e5-f852a8b958ea\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-llpfd" Oct 01 10:01:27 crc kubenswrapper[4787]: I1001 10:01:27.489790 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8dd2f969-9904-4848-b0e5-f852a8b958ea-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-llpfd\" (UID: \"8dd2f969-9904-4848-b0e5-f852a8b958ea\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-llpfd" Oct 01 10:01:27 crc kubenswrapper[4787]: I1001 10:01:27.489905 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8dd2f969-9904-4848-b0e5-f852a8b958ea-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-llpfd\" (UID: \"8dd2f969-9904-4848-b0e5-f852a8b958ea\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-llpfd" Oct 01 10:01:27 crc kubenswrapper[4787]: I1001 10:01:27.489971 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sw9z\" (UniqueName: \"kubernetes.io/projected/8dd2f969-9904-4848-b0e5-f852a8b958ea-kube-api-access-5sw9z\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-llpfd\" (UID: \"8dd2f969-9904-4848-b0e5-f852a8b958ea\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-llpfd" Oct 01 10:01:27 crc kubenswrapper[4787]: I1001 10:01:27.494889 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8dd2f969-9904-4848-b0e5-f852a8b958ea-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-llpfd\" (UID: \"8dd2f969-9904-4848-b0e5-f852a8b958ea\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-llpfd" Oct 01 10:01:27 crc kubenswrapper[4787]: I1001 10:01:27.503692 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8dd2f969-9904-4848-b0e5-f852a8b958ea-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-llpfd\" (UID: \"8dd2f969-9904-4848-b0e5-f852a8b958ea\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-llpfd" Oct 01 10:01:27 crc kubenswrapper[4787]: I1001 10:01:27.513125 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sw9z\" (UniqueName: \"kubernetes.io/projected/8dd2f969-9904-4848-b0e5-f852a8b958ea-kube-api-access-5sw9z\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-llpfd\" (UID: \"8dd2f969-9904-4848-b0e5-f852a8b958ea\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-llpfd" Oct 01 10:01:27 crc kubenswrapper[4787]: I1001 10:01:27.572520 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-llpfd" Oct 01 10:01:28 crc kubenswrapper[4787]: I1001 10:01:28.062966 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-llpfd"] Oct 01 10:01:28 crc kubenswrapper[4787]: I1001 10:01:28.069833 4787 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 10:01:28 crc kubenswrapper[4787]: I1001 10:01:28.154476 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-llpfd" event={"ID":"8dd2f969-9904-4848-b0e5-f852a8b958ea","Type":"ContainerStarted","Data":"c7f5dbd34c288c9e885751a295da80e0420ac836a4d72492bbfeb5dfc98c5500"} Oct 01 10:01:29 crc kubenswrapper[4787]: I1001 10:01:29.169101 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-llpfd" event={"ID":"8dd2f969-9904-4848-b0e5-f852a8b958ea","Type":"ContainerStarted","Data":"229cd38bffc9c220ba460d3aca06522f825f2d273c26497fdfd806f06456403d"} Oct 01 10:01:29 crc kubenswrapper[4787]: I1001 10:01:29.197605 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-llpfd" podStartSLOduration=1.7348102170000002 podStartE2EDuration="2.197574949s" podCreationTimestamp="2025-10-01 10:01:27 +0000 UTC" firstStartedPulling="2025-10-01 10:01:28.06955222 +0000 UTC m=+1520.184696377" lastFinishedPulling="2025-10-01 10:01:28.532316952 +0000 UTC m=+1520.647461109" observedRunningTime="2025-10-01 10:01:29.187634548 +0000 UTC m=+1521.302778705" watchObservedRunningTime="2025-10-01 10:01:29.197574949 +0000 UTC m=+1521.312719136" Oct 01 10:01:30 crc kubenswrapper[4787]: I1001 10:01:30.635815 4787 scope.go:117] "RemoveContainer" containerID="a683ee182b63428f19a3d11e7554ca1a4e3f554728ce4dd7334e736c8d8dd617" Oct 01 10:01:30 crc kubenswrapper[4787]: I1001 10:01:30.668916 4787 scope.go:117] "RemoveContainer" containerID="3f08c2d08d5b370333d23fbf069dd296eb3b2296c94e98ef187f5b7c07f3ed23" Oct 01 10:01:34 crc kubenswrapper[4787]: I1001 10:01:34.031526 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-2j6gl"] Oct 01 10:01:34 crc kubenswrapper[4787]: I1001 10:01:34.039004 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-2j6gl"] Oct 01 10:01:34 crc kubenswrapper[4787]: I1001 10:01:34.540342 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="535755f5-70d6-4fa0-8be1-c8add5e1283c" path="/var/lib/kubelet/pods/535755f5-70d6-4fa0-8be1-c8add5e1283c/volumes" Oct 01 10:01:35 crc kubenswrapper[4787]: I1001 10:01:35.032963 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-27c6-account-create-hqf2p"] Oct 01 10:01:35 crc kubenswrapper[4787]: I1001 10:01:35.041529 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-7321-account-create-gvbhd"] Oct 01 10:01:35 crc kubenswrapper[4787]: I1001 10:01:35.050189 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-27c6-account-create-hqf2p"] Oct 01 10:01:35 crc kubenswrapper[4787]: I1001 10:01:35.058053 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-7321-account-create-gvbhd"] Oct 01 10:01:35 crc kubenswrapper[4787]: I1001 10:01:35.523851 4787 scope.go:117] "RemoveContainer" containerID="e01333312cf21546b537b606d139d2d34e67fa8885c08f9ceb1f4c790ae4fd3a" Oct 01 10:01:35 crc kubenswrapper[4787]: E1001 10:01:35.524228 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:01:36 crc kubenswrapper[4787]: I1001 10:01:36.540412 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a3b9c9c-d4f0-4f2e-b2f3-44689860ec29" path="/var/lib/kubelet/pods/7a3b9c9c-d4f0-4f2e-b2f3-44689860ec29/volumes" Oct 01 10:01:36 crc kubenswrapper[4787]: I1001 10:01:36.542560 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3cbab8d-b6e2-420a-8456-efef86923af5" path="/var/lib/kubelet/pods/c3cbab8d-b6e2-420a-8456-efef86923af5/volumes" Oct 01 10:01:44 crc kubenswrapper[4787]: I1001 10:01:44.044419 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-de61-account-create-hxhr8"] Oct 01 10:01:44 crc kubenswrapper[4787]: I1001 10:01:44.058024 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-de61-account-create-hxhr8"] Oct 01 10:01:44 crc kubenswrapper[4787]: I1001 10:01:44.537616 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d2457dd-a699-48b4-86d0-69505da5345c" path="/var/lib/kubelet/pods/1d2457dd-a699-48b4-86d0-69505da5345c/volumes" Oct 01 10:01:49 crc kubenswrapper[4787]: I1001 10:01:49.523450 4787 scope.go:117] "RemoveContainer" containerID="e01333312cf21546b537b606d139d2d34e67fa8885c08f9ceb1f4c790ae4fd3a" Oct 01 10:01:49 crc kubenswrapper[4787]: E1001 10:01:49.524324 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:01:56 crc kubenswrapper[4787]: I1001 10:01:56.049917 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-jbw9g"] Oct 01 10:01:56 crc kubenswrapper[4787]: I1001 10:01:56.062426 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-8d4s5"] Oct 01 10:01:56 crc kubenswrapper[4787]: I1001 10:01:56.071239 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-8d4s5"] Oct 01 10:01:56 crc kubenswrapper[4787]: I1001 10:01:56.080392 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-rblh2"] Oct 01 10:01:56 crc kubenswrapper[4787]: I1001 10:01:56.087764 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-jbw9g"] Oct 01 10:01:56 crc kubenswrapper[4787]: I1001 10:01:56.094896 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-rblh2"] Oct 01 10:01:56 crc kubenswrapper[4787]: I1001 10:01:56.534418 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a294c79-3cc4-4070-86c7-baad1226a08e" path="/var/lib/kubelet/pods/8a294c79-3cc4-4070-86c7-baad1226a08e/volumes" Oct 01 10:01:56 crc kubenswrapper[4787]: I1001 10:01:56.534936 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8de5a45d-bdad-4e9b-9385-13ac70b41009" path="/var/lib/kubelet/pods/8de5a45d-bdad-4e9b-9385-13ac70b41009/volumes" Oct 01 10:01:56 crc kubenswrapper[4787]: I1001 10:01:56.535424 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d39dce3-8e56-4d16-ad94-8f582723de4e" path="/var/lib/kubelet/pods/9d39dce3-8e56-4d16-ad94-8f582723de4e/volumes" Oct 01 10:01:59 crc kubenswrapper[4787]: I1001 10:01:59.031056 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-kckrj"] Oct 01 10:01:59 crc kubenswrapper[4787]: I1001 10:01:59.039030 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-kckrj"] Oct 01 10:02:00 crc kubenswrapper[4787]: I1001 10:02:00.031272 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-pd2hv"] Oct 01 10:02:00 crc kubenswrapper[4787]: I1001 10:02:00.041345 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-pd2hv"] Oct 01 10:02:00 crc kubenswrapper[4787]: I1001 10:02:00.535823 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5fd5149-1773-49d8-8612-fe5eb43ddda8" path="/var/lib/kubelet/pods/c5fd5149-1773-49d8-8612-fe5eb43ddda8/volumes" Oct 01 10:02:00 crc kubenswrapper[4787]: I1001 10:02:00.536869 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7251286-7de7-48ef-a418-6d67749e40b8" path="/var/lib/kubelet/pods/f7251286-7de7-48ef-a418-6d67749e40b8/volumes" Oct 01 10:02:04 crc kubenswrapper[4787]: I1001 10:02:04.524188 4787 scope.go:117] "RemoveContainer" containerID="e01333312cf21546b537b606d139d2d34e67fa8885c08f9ceb1f4c790ae4fd3a" Oct 01 10:02:04 crc kubenswrapper[4787]: E1001 10:02:04.525124 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:02:15 crc kubenswrapper[4787]: I1001 10:02:15.524171 4787 scope.go:117] "RemoveContainer" containerID="e01333312cf21546b537b606d139d2d34e67fa8885c08f9ceb1f4c790ae4fd3a" Oct 01 10:02:15 crc kubenswrapper[4787]: E1001 10:02:15.524953 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:02:26 crc kubenswrapper[4787]: I1001 10:02:26.046237 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-cf80-account-create-zz7cn"] Oct 01 10:02:26 crc kubenswrapper[4787]: I1001 10:02:26.056958 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-819c-account-create-86wn2"] Oct 01 10:02:26 crc kubenswrapper[4787]: I1001 10:02:26.066471 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-819c-account-create-86wn2"] Oct 01 10:02:26 crc kubenswrapper[4787]: I1001 10:02:26.076566 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-cf80-account-create-zz7cn"] Oct 01 10:02:26 crc kubenswrapper[4787]: I1001 10:02:26.536183 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22964b11-e82b-4dfa-ac42-70005cbb5caf" path="/var/lib/kubelet/pods/22964b11-e82b-4dfa-ac42-70005cbb5caf/volumes" Oct 01 10:02:26 crc kubenswrapper[4787]: I1001 10:02:26.536801 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54517204-c321-4959-851b-9c642f07a7a7" path="/var/lib/kubelet/pods/54517204-c321-4959-851b-9c642f07a7a7/volumes" Oct 01 10:02:27 crc kubenswrapper[4787]: I1001 10:02:27.023733 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-eafd-account-create-h97jm"] Oct 01 10:02:27 crc kubenswrapper[4787]: I1001 10:02:27.030983 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-eafd-account-create-h97jm"] Oct 01 10:02:28 crc kubenswrapper[4787]: I1001 10:02:28.023034 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-slvbz"] Oct 01 10:02:28 crc kubenswrapper[4787]: I1001 10:02:28.030250 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-slvbz"] Oct 01 10:02:28 crc kubenswrapper[4787]: I1001 10:02:28.537919 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6af63266-547b-4537-9290-338b2c0a2d73" path="/var/lib/kubelet/pods/6af63266-547b-4537-9290-338b2c0a2d73/volumes" Oct 01 10:02:28 crc kubenswrapper[4787]: I1001 10:02:28.538681 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcadc39c-6259-49a9-9f5f-38a80545b5a4" path="/var/lib/kubelet/pods/bcadc39c-6259-49a9-9f5f-38a80545b5a4/volumes" Oct 01 10:02:29 crc kubenswrapper[4787]: I1001 10:02:29.525096 4787 scope.go:117] "RemoveContainer" containerID="e01333312cf21546b537b606d139d2d34e67fa8885c08f9ceb1f4c790ae4fd3a" Oct 01 10:02:29 crc kubenswrapper[4787]: E1001 10:02:29.525575 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:02:30 crc kubenswrapper[4787]: I1001 10:02:30.773682 4787 scope.go:117] "RemoveContainer" containerID="65b2293dbb8338d9092baa098fc300c3527196b0d2d7eaaf08fdb65b4dea50b7" Oct 01 10:02:30 crc kubenswrapper[4787]: I1001 10:02:30.808169 4787 scope.go:117] "RemoveContainer" containerID="54dc15565137169bed13359225ad074892e4991a14fa5bee94ec079f11e584e9" Oct 01 10:02:30 crc kubenswrapper[4787]: I1001 10:02:30.876577 4787 scope.go:117] "RemoveContainer" containerID="81db296f8f18bdde1e19f761178b08968998618eefe18e165ce27d4659f5cfd9" Oct 01 10:02:30 crc kubenswrapper[4787]: I1001 10:02:30.911582 4787 scope.go:117] "RemoveContainer" containerID="72eab7966f013f105d29b360d7b6c98c01ddfc7ab40598ddead1f1c8e77924f2" Oct 01 10:02:30 crc kubenswrapper[4787]: I1001 10:02:30.970920 4787 scope.go:117] "RemoveContainer" containerID="b80433841686fb3ebfee3df8efba794169f6019700338907134cc87d5328b4e4" Oct 01 10:02:31 crc kubenswrapper[4787]: I1001 10:02:31.020107 4787 scope.go:117] "RemoveContainer" containerID="615c6ef799483af41a0c23e5042cd9bad9e09072c68ddd6ea5acb03343c9db67" Oct 01 10:02:31 crc kubenswrapper[4787]: I1001 10:02:31.053683 4787 scope.go:117] "RemoveContainer" containerID="ffbc1e8fe63ccb959ba0dd0a30701a2b032acb532b2d1bdb5ffa014ef6d9b226" Oct 01 10:02:31 crc kubenswrapper[4787]: I1001 10:02:31.086386 4787 scope.go:117] "RemoveContainer" containerID="13664b5c7d47d027839c3d62ffea9e5f2cada72768847edbf9afb9e57d3c6c80" Oct 01 10:02:31 crc kubenswrapper[4787]: I1001 10:02:31.106120 4787 scope.go:117] "RemoveContainer" containerID="dfc015bb702961c5fad27e0d1f72d6de0ae82cd36b5eece674ed64f963d1e937" Oct 01 10:02:31 crc kubenswrapper[4787]: I1001 10:02:31.124129 4787 scope.go:117] "RemoveContainer" containerID="44d649cfd48ca68d099ccc760ec2c638396f532602b5f8d951457214115f5bd7" Oct 01 10:02:31 crc kubenswrapper[4787]: I1001 10:02:31.142197 4787 scope.go:117] "RemoveContainer" containerID="9615249f72189c58f8030de6ab45f6725cd8c892d5b91b1d6092b363c976998d" Oct 01 10:02:31 crc kubenswrapper[4787]: I1001 10:02:31.162315 4787 scope.go:117] "RemoveContainer" containerID="36087873618e96f830b48167fbcbb872b6a1cae4fbd26be0edf8d3a2923a3560" Oct 01 10:02:31 crc kubenswrapper[4787]: I1001 10:02:31.189280 4787 scope.go:117] "RemoveContainer" containerID="b46dfa12aa0b9ae2977caa8b3df6e30cf29ff756fb395e9e390f056bcced7b29" Oct 01 10:02:33 crc kubenswrapper[4787]: I1001 10:02:33.037385 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-dx2gc"] Oct 01 10:02:33 crc kubenswrapper[4787]: I1001 10:02:33.046618 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-dx2gc"] Oct 01 10:02:34 crc kubenswrapper[4787]: I1001 10:02:34.538191 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="345ecb87-2ef8-4829-8137-156dec6d0e80" path="/var/lib/kubelet/pods/345ecb87-2ef8-4829-8137-156dec6d0e80/volumes" Oct 01 10:02:39 crc kubenswrapper[4787]: E1001 10:02:39.515732 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8dd2f969_9904_4848_b0e5_f852a8b958ea.slice/crio-229cd38bffc9c220ba460d3aca06522f825f2d273c26497fdfd806f06456403d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8dd2f969_9904_4848_b0e5_f852a8b958ea.slice/crio-conmon-229cd38bffc9c220ba460d3aca06522f825f2d273c26497fdfd806f06456403d.scope\": RecentStats: unable to find data in memory cache]" Oct 01 10:02:39 crc kubenswrapper[4787]: I1001 10:02:39.929915 4787 generic.go:334] "Generic (PLEG): container finished" podID="8dd2f969-9904-4848-b0e5-f852a8b958ea" containerID="229cd38bffc9c220ba460d3aca06522f825f2d273c26497fdfd806f06456403d" exitCode=0 Oct 01 10:02:39 crc kubenswrapper[4787]: I1001 10:02:39.929951 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-llpfd" event={"ID":"8dd2f969-9904-4848-b0e5-f852a8b958ea","Type":"ContainerDied","Data":"229cd38bffc9c220ba460d3aca06522f825f2d273c26497fdfd806f06456403d"} Oct 01 10:02:41 crc kubenswrapper[4787]: I1001 10:02:41.388166 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-llpfd" Oct 01 10:02:41 crc kubenswrapper[4787]: I1001 10:02:41.524018 4787 scope.go:117] "RemoveContainer" containerID="e01333312cf21546b537b606d139d2d34e67fa8885c08f9ceb1f4c790ae4fd3a" Oct 01 10:02:41 crc kubenswrapper[4787]: E1001 10:02:41.525859 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:02:41 crc kubenswrapper[4787]: I1001 10:02:41.571144 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8dd2f969-9904-4848-b0e5-f852a8b958ea-inventory\") pod \"8dd2f969-9904-4848-b0e5-f852a8b958ea\" (UID: \"8dd2f969-9904-4848-b0e5-f852a8b958ea\") " Oct 01 10:02:41 crc kubenswrapper[4787]: I1001 10:02:41.571255 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8dd2f969-9904-4848-b0e5-f852a8b958ea-ssh-key\") pod \"8dd2f969-9904-4848-b0e5-f852a8b958ea\" (UID: \"8dd2f969-9904-4848-b0e5-f852a8b958ea\") " Oct 01 10:02:41 crc kubenswrapper[4787]: I1001 10:02:41.571505 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5sw9z\" (UniqueName: \"kubernetes.io/projected/8dd2f969-9904-4848-b0e5-f852a8b958ea-kube-api-access-5sw9z\") pod \"8dd2f969-9904-4848-b0e5-f852a8b958ea\" (UID: \"8dd2f969-9904-4848-b0e5-f852a8b958ea\") " Oct 01 10:02:41 crc kubenswrapper[4787]: I1001 10:02:41.579295 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dd2f969-9904-4848-b0e5-f852a8b958ea-kube-api-access-5sw9z" (OuterVolumeSpecName: "kube-api-access-5sw9z") pod "8dd2f969-9904-4848-b0e5-f852a8b958ea" (UID: "8dd2f969-9904-4848-b0e5-f852a8b958ea"). InnerVolumeSpecName "kube-api-access-5sw9z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:02:41 crc kubenswrapper[4787]: I1001 10:02:41.600731 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dd2f969-9904-4848-b0e5-f852a8b958ea-inventory" (OuterVolumeSpecName: "inventory") pod "8dd2f969-9904-4848-b0e5-f852a8b958ea" (UID: "8dd2f969-9904-4848-b0e5-f852a8b958ea"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:02:41 crc kubenswrapper[4787]: I1001 10:02:41.609625 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dd2f969-9904-4848-b0e5-f852a8b958ea-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8dd2f969-9904-4848-b0e5-f852a8b958ea" (UID: "8dd2f969-9904-4848-b0e5-f852a8b958ea"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:02:41 crc kubenswrapper[4787]: I1001 10:02:41.674325 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5sw9z\" (UniqueName: \"kubernetes.io/projected/8dd2f969-9904-4848-b0e5-f852a8b958ea-kube-api-access-5sw9z\") on node \"crc\" DevicePath \"\"" Oct 01 10:02:41 crc kubenswrapper[4787]: I1001 10:02:41.674387 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8dd2f969-9904-4848-b0e5-f852a8b958ea-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 10:02:41 crc kubenswrapper[4787]: I1001 10:02:41.674414 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8dd2f969-9904-4848-b0e5-f852a8b958ea-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 10:02:41 crc kubenswrapper[4787]: I1001 10:02:41.950986 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-llpfd" event={"ID":"8dd2f969-9904-4848-b0e5-f852a8b958ea","Type":"ContainerDied","Data":"c7f5dbd34c288c9e885751a295da80e0420ac836a4d72492bbfeb5dfc98c5500"} Oct 01 10:02:41 crc kubenswrapper[4787]: I1001 10:02:41.951051 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7f5dbd34c288c9e885751a295da80e0420ac836a4d72492bbfeb5dfc98c5500" Oct 01 10:02:41 crc kubenswrapper[4787]: I1001 10:02:41.951175 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-llpfd" Oct 01 10:02:42 crc kubenswrapper[4787]: I1001 10:02:42.040446 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nkbqd"] Oct 01 10:02:42 crc kubenswrapper[4787]: E1001 10:02:42.040971 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dd2f969-9904-4848-b0e5-f852a8b958ea" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 01 10:02:42 crc kubenswrapper[4787]: I1001 10:02:42.040995 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dd2f969-9904-4848-b0e5-f852a8b958ea" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 01 10:02:42 crc kubenswrapper[4787]: I1001 10:02:42.041300 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dd2f969-9904-4848-b0e5-f852a8b958ea" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 01 10:02:42 crc kubenswrapper[4787]: I1001 10:02:42.042106 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nkbqd" Oct 01 10:02:42 crc kubenswrapper[4787]: I1001 10:02:42.044224 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 10:02:42 crc kubenswrapper[4787]: I1001 10:02:42.044675 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 10:02:42 crc kubenswrapper[4787]: I1001 10:02:42.045768 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 10:02:42 crc kubenswrapper[4787]: I1001 10:02:42.046861 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vh5v6" Oct 01 10:02:42 crc kubenswrapper[4787]: I1001 10:02:42.047939 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nkbqd"] Oct 01 10:02:42 crc kubenswrapper[4787]: I1001 10:02:42.186995 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08369955-6564-47c7-bed0-64893e898e1b-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-nkbqd\" (UID: \"08369955-6564-47c7-bed0-64893e898e1b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nkbqd" Oct 01 10:02:42 crc kubenswrapper[4787]: I1001 10:02:42.187366 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wc78c\" (UniqueName: \"kubernetes.io/projected/08369955-6564-47c7-bed0-64893e898e1b-kube-api-access-wc78c\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-nkbqd\" (UID: \"08369955-6564-47c7-bed0-64893e898e1b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nkbqd" Oct 01 10:02:42 crc kubenswrapper[4787]: I1001 10:02:42.187577 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/08369955-6564-47c7-bed0-64893e898e1b-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-nkbqd\" (UID: \"08369955-6564-47c7-bed0-64893e898e1b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nkbqd" Oct 01 10:02:42 crc kubenswrapper[4787]: I1001 10:02:42.289220 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08369955-6564-47c7-bed0-64893e898e1b-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-nkbqd\" (UID: \"08369955-6564-47c7-bed0-64893e898e1b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nkbqd" Oct 01 10:02:42 crc kubenswrapper[4787]: I1001 10:02:42.289351 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wc78c\" (UniqueName: \"kubernetes.io/projected/08369955-6564-47c7-bed0-64893e898e1b-kube-api-access-wc78c\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-nkbqd\" (UID: \"08369955-6564-47c7-bed0-64893e898e1b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nkbqd" Oct 01 10:02:42 crc kubenswrapper[4787]: I1001 10:02:42.289381 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/08369955-6564-47c7-bed0-64893e898e1b-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-nkbqd\" (UID: \"08369955-6564-47c7-bed0-64893e898e1b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nkbqd" Oct 01 10:02:42 crc kubenswrapper[4787]: I1001 10:02:42.294870 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/08369955-6564-47c7-bed0-64893e898e1b-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-nkbqd\" (UID: \"08369955-6564-47c7-bed0-64893e898e1b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nkbqd" Oct 01 10:02:42 crc kubenswrapper[4787]: I1001 10:02:42.295845 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08369955-6564-47c7-bed0-64893e898e1b-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-nkbqd\" (UID: \"08369955-6564-47c7-bed0-64893e898e1b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nkbqd" Oct 01 10:02:42 crc kubenswrapper[4787]: I1001 10:02:42.321407 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wc78c\" (UniqueName: \"kubernetes.io/projected/08369955-6564-47c7-bed0-64893e898e1b-kube-api-access-wc78c\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-nkbqd\" (UID: \"08369955-6564-47c7-bed0-64893e898e1b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nkbqd" Oct 01 10:02:42 crc kubenswrapper[4787]: I1001 10:02:42.363244 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nkbqd" Oct 01 10:02:42 crc kubenswrapper[4787]: I1001 10:02:42.818492 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nkbqd"] Oct 01 10:02:42 crc kubenswrapper[4787]: I1001 10:02:42.959293 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nkbqd" event={"ID":"08369955-6564-47c7-bed0-64893e898e1b","Type":"ContainerStarted","Data":"84f18de923f5cc4c94795613bfe34db26bf50de213c9ae527ec6699e215f4475"} Oct 01 10:02:43 crc kubenswrapper[4787]: I1001 10:02:43.971544 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nkbqd" event={"ID":"08369955-6564-47c7-bed0-64893e898e1b","Type":"ContainerStarted","Data":"7498cb84cbd719e1ec616fe5bbeee9b0bd9f8a76724426ee14b5aa552e9a3b6a"} Oct 01 10:02:43 crc kubenswrapper[4787]: I1001 10:02:43.997531 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nkbqd" podStartSLOduration=1.570031786 podStartE2EDuration="1.997508441s" podCreationTimestamp="2025-10-01 10:02:42 +0000 UTC" firstStartedPulling="2025-10-01 10:02:42.816190099 +0000 UTC m=+1594.931334256" lastFinishedPulling="2025-10-01 10:02:43.243666754 +0000 UTC m=+1595.358810911" observedRunningTime="2025-10-01 10:02:43.99048188 +0000 UTC m=+1596.105626037" watchObservedRunningTime="2025-10-01 10:02:43.997508441 +0000 UTC m=+1596.112652598" Oct 01 10:02:49 crc kubenswrapper[4787]: I1001 10:02:49.025549 4787 generic.go:334] "Generic (PLEG): container finished" podID="08369955-6564-47c7-bed0-64893e898e1b" containerID="7498cb84cbd719e1ec616fe5bbeee9b0bd9f8a76724426ee14b5aa552e9a3b6a" exitCode=0 Oct 01 10:02:49 crc kubenswrapper[4787]: I1001 10:02:49.025672 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nkbqd" event={"ID":"08369955-6564-47c7-bed0-64893e898e1b","Type":"ContainerDied","Data":"7498cb84cbd719e1ec616fe5bbeee9b0bd9f8a76724426ee14b5aa552e9a3b6a"} Oct 01 10:02:50 crc kubenswrapper[4787]: I1001 10:02:50.443673 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nkbqd" Oct 01 10:02:50 crc kubenswrapper[4787]: I1001 10:02:50.587101 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08369955-6564-47c7-bed0-64893e898e1b-ssh-key\") pod \"08369955-6564-47c7-bed0-64893e898e1b\" (UID: \"08369955-6564-47c7-bed0-64893e898e1b\") " Oct 01 10:02:50 crc kubenswrapper[4787]: I1001 10:02:50.587256 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/08369955-6564-47c7-bed0-64893e898e1b-inventory\") pod \"08369955-6564-47c7-bed0-64893e898e1b\" (UID: \"08369955-6564-47c7-bed0-64893e898e1b\") " Oct 01 10:02:50 crc kubenswrapper[4787]: I1001 10:02:50.588668 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wc78c\" (UniqueName: \"kubernetes.io/projected/08369955-6564-47c7-bed0-64893e898e1b-kube-api-access-wc78c\") pod \"08369955-6564-47c7-bed0-64893e898e1b\" (UID: \"08369955-6564-47c7-bed0-64893e898e1b\") " Oct 01 10:02:50 crc kubenswrapper[4787]: I1001 10:02:50.607422 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08369955-6564-47c7-bed0-64893e898e1b-kube-api-access-wc78c" (OuterVolumeSpecName: "kube-api-access-wc78c") pod "08369955-6564-47c7-bed0-64893e898e1b" (UID: "08369955-6564-47c7-bed0-64893e898e1b"). InnerVolumeSpecName "kube-api-access-wc78c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:02:50 crc kubenswrapper[4787]: I1001 10:02:50.629021 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08369955-6564-47c7-bed0-64893e898e1b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "08369955-6564-47c7-bed0-64893e898e1b" (UID: "08369955-6564-47c7-bed0-64893e898e1b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:02:50 crc kubenswrapper[4787]: I1001 10:02:50.632373 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08369955-6564-47c7-bed0-64893e898e1b-inventory" (OuterVolumeSpecName: "inventory") pod "08369955-6564-47c7-bed0-64893e898e1b" (UID: "08369955-6564-47c7-bed0-64893e898e1b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:02:50 crc kubenswrapper[4787]: I1001 10:02:50.691184 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08369955-6564-47c7-bed0-64893e898e1b-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 10:02:50 crc kubenswrapper[4787]: I1001 10:02:50.691216 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/08369955-6564-47c7-bed0-64893e898e1b-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 10:02:50 crc kubenswrapper[4787]: I1001 10:02:50.691226 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wc78c\" (UniqueName: \"kubernetes.io/projected/08369955-6564-47c7-bed0-64893e898e1b-kube-api-access-wc78c\") on node \"crc\" DevicePath \"\"" Oct 01 10:02:51 crc kubenswrapper[4787]: I1001 10:02:51.047722 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nkbqd" event={"ID":"08369955-6564-47c7-bed0-64893e898e1b","Type":"ContainerDied","Data":"84f18de923f5cc4c94795613bfe34db26bf50de213c9ae527ec6699e215f4475"} Oct 01 10:02:51 crc kubenswrapper[4787]: I1001 10:02:51.047766 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="84f18de923f5cc4c94795613bfe34db26bf50de213c9ae527ec6699e215f4475" Oct 01 10:02:51 crc kubenswrapper[4787]: I1001 10:02:51.047785 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nkbqd" Oct 01 10:02:51 crc kubenswrapper[4787]: I1001 10:02:51.120138 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-qvjjs"] Oct 01 10:02:51 crc kubenswrapper[4787]: E1001 10:02:51.120651 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08369955-6564-47c7-bed0-64893e898e1b" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 01 10:02:51 crc kubenswrapper[4787]: I1001 10:02:51.120678 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="08369955-6564-47c7-bed0-64893e898e1b" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 01 10:02:51 crc kubenswrapper[4787]: I1001 10:02:51.120942 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="08369955-6564-47c7-bed0-64893e898e1b" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 01 10:02:51 crc kubenswrapper[4787]: I1001 10:02:51.121764 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qvjjs" Oct 01 10:02:51 crc kubenswrapper[4787]: I1001 10:02:51.127128 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 10:02:51 crc kubenswrapper[4787]: I1001 10:02:51.127390 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 10:02:51 crc kubenswrapper[4787]: I1001 10:02:51.127529 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 10:02:51 crc kubenswrapper[4787]: I1001 10:02:51.135425 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vh5v6" Oct 01 10:02:51 crc kubenswrapper[4787]: I1001 10:02:51.137294 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-qvjjs"] Oct 01 10:02:51 crc kubenswrapper[4787]: I1001 10:02:51.303436 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbzr7\" (UniqueName: \"kubernetes.io/projected/654cfd15-20ca-4754-9de7-ebe67905c8d2-kube-api-access-rbzr7\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qvjjs\" (UID: \"654cfd15-20ca-4754-9de7-ebe67905c8d2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qvjjs" Oct 01 10:02:51 crc kubenswrapper[4787]: I1001 10:02:51.303504 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/654cfd15-20ca-4754-9de7-ebe67905c8d2-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qvjjs\" (UID: \"654cfd15-20ca-4754-9de7-ebe67905c8d2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qvjjs" Oct 01 10:02:51 crc kubenswrapper[4787]: I1001 10:02:51.303535 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/654cfd15-20ca-4754-9de7-ebe67905c8d2-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qvjjs\" (UID: \"654cfd15-20ca-4754-9de7-ebe67905c8d2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qvjjs" Oct 01 10:02:51 crc kubenswrapper[4787]: I1001 10:02:51.406304 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbzr7\" (UniqueName: \"kubernetes.io/projected/654cfd15-20ca-4754-9de7-ebe67905c8d2-kube-api-access-rbzr7\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qvjjs\" (UID: \"654cfd15-20ca-4754-9de7-ebe67905c8d2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qvjjs" Oct 01 10:02:51 crc kubenswrapper[4787]: I1001 10:02:51.406374 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/654cfd15-20ca-4754-9de7-ebe67905c8d2-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qvjjs\" (UID: \"654cfd15-20ca-4754-9de7-ebe67905c8d2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qvjjs" Oct 01 10:02:51 crc kubenswrapper[4787]: I1001 10:02:51.406394 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/654cfd15-20ca-4754-9de7-ebe67905c8d2-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qvjjs\" (UID: \"654cfd15-20ca-4754-9de7-ebe67905c8d2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qvjjs" Oct 01 10:02:51 crc kubenswrapper[4787]: I1001 10:02:51.410759 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/654cfd15-20ca-4754-9de7-ebe67905c8d2-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qvjjs\" (UID: \"654cfd15-20ca-4754-9de7-ebe67905c8d2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qvjjs" Oct 01 10:02:51 crc kubenswrapper[4787]: I1001 10:02:51.411173 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/654cfd15-20ca-4754-9de7-ebe67905c8d2-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qvjjs\" (UID: \"654cfd15-20ca-4754-9de7-ebe67905c8d2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qvjjs" Oct 01 10:02:51 crc kubenswrapper[4787]: I1001 10:02:51.438230 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbzr7\" (UniqueName: \"kubernetes.io/projected/654cfd15-20ca-4754-9de7-ebe67905c8d2-kube-api-access-rbzr7\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qvjjs\" (UID: \"654cfd15-20ca-4754-9de7-ebe67905c8d2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qvjjs" Oct 01 10:02:51 crc kubenswrapper[4787]: I1001 10:02:51.446820 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qvjjs" Oct 01 10:02:51 crc kubenswrapper[4787]: I1001 10:02:51.990682 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-qvjjs"] Oct 01 10:02:51 crc kubenswrapper[4787]: W1001 10:02:51.999745 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod654cfd15_20ca_4754_9de7_ebe67905c8d2.slice/crio-f852176119d6a3ba8c9b52a487271ccd472de6cf66edd0a20cba0b849f737ada WatchSource:0}: Error finding container f852176119d6a3ba8c9b52a487271ccd472de6cf66edd0a20cba0b849f737ada: Status 404 returned error can't find the container with id f852176119d6a3ba8c9b52a487271ccd472de6cf66edd0a20cba0b849f737ada Oct 01 10:02:52 crc kubenswrapper[4787]: I1001 10:02:52.058302 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qvjjs" event={"ID":"654cfd15-20ca-4754-9de7-ebe67905c8d2","Type":"ContainerStarted","Data":"f852176119d6a3ba8c9b52a487271ccd472de6cf66edd0a20cba0b849f737ada"} Oct 01 10:02:53 crc kubenswrapper[4787]: I1001 10:02:53.067888 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qvjjs" event={"ID":"654cfd15-20ca-4754-9de7-ebe67905c8d2","Type":"ContainerStarted","Data":"e5461f6310e0268258f39b4a00e685ab23d8a1c06d3f87b09e359d256279077d"} Oct 01 10:02:53 crc kubenswrapper[4787]: I1001 10:02:53.087191 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qvjjs" podStartSLOduration=1.6298271199999999 podStartE2EDuration="2.087175151s" podCreationTimestamp="2025-10-01 10:02:51 +0000 UTC" firstStartedPulling="2025-10-01 10:02:52.003282053 +0000 UTC m=+1604.118426210" lastFinishedPulling="2025-10-01 10:02:52.460630084 +0000 UTC m=+1604.575774241" observedRunningTime="2025-10-01 10:02:53.085881209 +0000 UTC m=+1605.201025376" watchObservedRunningTime="2025-10-01 10:02:53.087175151 +0000 UTC m=+1605.202319308" Oct 01 10:02:54 crc kubenswrapper[4787]: I1001 10:02:54.524267 4787 scope.go:117] "RemoveContainer" containerID="e01333312cf21546b537b606d139d2d34e67fa8885c08f9ceb1f4c790ae4fd3a" Oct 01 10:02:54 crc kubenswrapper[4787]: E1001 10:02:54.525155 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:02:56 crc kubenswrapper[4787]: I1001 10:02:56.066054 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-kvcbj"] Oct 01 10:02:56 crc kubenswrapper[4787]: I1001 10:02:56.077642 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-kvcbj"] Oct 01 10:02:56 crc kubenswrapper[4787]: I1001 10:02:56.537134 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18f38986-fd49-4f82-ad99-0a73264877a6" path="/var/lib/kubelet/pods/18f38986-fd49-4f82-ad99-0a73264877a6/volumes" Oct 01 10:03:02 crc kubenswrapper[4787]: I1001 10:03:02.031399 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-wndd6"] Oct 01 10:03:02 crc kubenswrapper[4787]: I1001 10:03:02.038457 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-wndd6"] Oct 01 10:03:02 crc kubenswrapper[4787]: I1001 10:03:02.535048 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15111beb-b05b-440f-a4c9-077eca1c37d1" path="/var/lib/kubelet/pods/15111beb-b05b-440f-a4c9-077eca1c37d1/volumes" Oct 01 10:03:08 crc kubenswrapper[4787]: I1001 10:03:08.530488 4787 scope.go:117] "RemoveContainer" containerID="e01333312cf21546b537b606d139d2d34e67fa8885c08f9ceb1f4c790ae4fd3a" Oct 01 10:03:08 crc kubenswrapper[4787]: E1001 10:03:08.531823 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:03:21 crc kubenswrapper[4787]: I1001 10:03:21.524133 4787 scope.go:117] "RemoveContainer" containerID="e01333312cf21546b537b606d139d2d34e67fa8885c08f9ceb1f4c790ae4fd3a" Oct 01 10:03:21 crc kubenswrapper[4787]: E1001 10:03:21.525266 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:03:22 crc kubenswrapper[4787]: I1001 10:03:22.040203 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-8mrcw"] Oct 01 10:03:22 crc kubenswrapper[4787]: I1001 10:03:22.048672 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-8mrcw"] Oct 01 10:03:22 crc kubenswrapper[4787]: I1001 10:03:22.535131 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca8b9373-e5e5-4dc1-8995-44ae47abbb13" path="/var/lib/kubelet/pods/ca8b9373-e5e5-4dc1-8995-44ae47abbb13/volumes" Oct 01 10:03:25 crc kubenswrapper[4787]: I1001 10:03:25.045132 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-9kp59"] Oct 01 10:03:25 crc kubenswrapper[4787]: I1001 10:03:25.052500 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-69qw2"] Oct 01 10:03:25 crc kubenswrapper[4787]: I1001 10:03:25.065016 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-69qw2"] Oct 01 10:03:25 crc kubenswrapper[4787]: I1001 10:03:25.076940 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-9kp59"] Oct 01 10:03:26 crc kubenswrapper[4787]: I1001 10:03:26.031006 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-nhmnw"] Oct 01 10:03:26 crc kubenswrapper[4787]: I1001 10:03:26.043990 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-nhmnw"] Oct 01 10:03:26 crc kubenswrapper[4787]: I1001 10:03:26.546933 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66351c20-c4db-489c-857f-c82107d2be61" path="/var/lib/kubelet/pods/66351c20-c4db-489c-857f-c82107d2be61/volumes" Oct 01 10:03:26 crc kubenswrapper[4787]: I1001 10:03:26.551807 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="931c8c97-57b0-4b0c-a97d-498eaad48543" path="/var/lib/kubelet/pods/931c8c97-57b0-4b0c-a97d-498eaad48543/volumes" Oct 01 10:03:26 crc kubenswrapper[4787]: I1001 10:03:26.553108 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7ba3660-dca6-4843-b560-efd0ea158258" path="/var/lib/kubelet/pods/a7ba3660-dca6-4843-b560-efd0ea158258/volumes" Oct 01 10:03:27 crc kubenswrapper[4787]: I1001 10:03:27.393923 4787 generic.go:334] "Generic (PLEG): container finished" podID="654cfd15-20ca-4754-9de7-ebe67905c8d2" containerID="e5461f6310e0268258f39b4a00e685ab23d8a1c06d3f87b09e359d256279077d" exitCode=0 Oct 01 10:03:27 crc kubenswrapper[4787]: I1001 10:03:27.393966 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qvjjs" event={"ID":"654cfd15-20ca-4754-9de7-ebe67905c8d2","Type":"ContainerDied","Data":"e5461f6310e0268258f39b4a00e685ab23d8a1c06d3f87b09e359d256279077d"} Oct 01 10:03:28 crc kubenswrapper[4787]: I1001 10:03:28.826507 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qvjjs" Oct 01 10:03:28 crc kubenswrapper[4787]: I1001 10:03:28.900547 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbzr7\" (UniqueName: \"kubernetes.io/projected/654cfd15-20ca-4754-9de7-ebe67905c8d2-kube-api-access-rbzr7\") pod \"654cfd15-20ca-4754-9de7-ebe67905c8d2\" (UID: \"654cfd15-20ca-4754-9de7-ebe67905c8d2\") " Oct 01 10:03:28 crc kubenswrapper[4787]: I1001 10:03:28.900634 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/654cfd15-20ca-4754-9de7-ebe67905c8d2-ssh-key\") pod \"654cfd15-20ca-4754-9de7-ebe67905c8d2\" (UID: \"654cfd15-20ca-4754-9de7-ebe67905c8d2\") " Oct 01 10:03:28 crc kubenswrapper[4787]: I1001 10:03:28.900752 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/654cfd15-20ca-4754-9de7-ebe67905c8d2-inventory\") pod \"654cfd15-20ca-4754-9de7-ebe67905c8d2\" (UID: \"654cfd15-20ca-4754-9de7-ebe67905c8d2\") " Oct 01 10:03:28 crc kubenswrapper[4787]: I1001 10:03:28.908592 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/654cfd15-20ca-4754-9de7-ebe67905c8d2-kube-api-access-rbzr7" (OuterVolumeSpecName: "kube-api-access-rbzr7") pod "654cfd15-20ca-4754-9de7-ebe67905c8d2" (UID: "654cfd15-20ca-4754-9de7-ebe67905c8d2"). InnerVolumeSpecName "kube-api-access-rbzr7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:03:28 crc kubenswrapper[4787]: I1001 10:03:28.929892 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/654cfd15-20ca-4754-9de7-ebe67905c8d2-inventory" (OuterVolumeSpecName: "inventory") pod "654cfd15-20ca-4754-9de7-ebe67905c8d2" (UID: "654cfd15-20ca-4754-9de7-ebe67905c8d2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:03:28 crc kubenswrapper[4787]: I1001 10:03:28.931377 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/654cfd15-20ca-4754-9de7-ebe67905c8d2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "654cfd15-20ca-4754-9de7-ebe67905c8d2" (UID: "654cfd15-20ca-4754-9de7-ebe67905c8d2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:03:29 crc kubenswrapper[4787]: I1001 10:03:29.003848 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/654cfd15-20ca-4754-9de7-ebe67905c8d2-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 10:03:29 crc kubenswrapper[4787]: I1001 10:03:29.003901 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbzr7\" (UniqueName: \"kubernetes.io/projected/654cfd15-20ca-4754-9de7-ebe67905c8d2-kube-api-access-rbzr7\") on node \"crc\" DevicePath \"\"" Oct 01 10:03:29 crc kubenswrapper[4787]: I1001 10:03:29.003920 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/654cfd15-20ca-4754-9de7-ebe67905c8d2-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 10:03:29 crc kubenswrapper[4787]: I1001 10:03:29.413619 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qvjjs" event={"ID":"654cfd15-20ca-4754-9de7-ebe67905c8d2","Type":"ContainerDied","Data":"f852176119d6a3ba8c9b52a487271ccd472de6cf66edd0a20cba0b849f737ada"} Oct 01 10:03:29 crc kubenswrapper[4787]: I1001 10:03:29.413677 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f852176119d6a3ba8c9b52a487271ccd472de6cf66edd0a20cba0b849f737ada" Oct 01 10:03:29 crc kubenswrapper[4787]: I1001 10:03:29.413980 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qvjjs" Oct 01 10:03:29 crc kubenswrapper[4787]: I1001 10:03:29.501837 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zwc7x"] Oct 01 10:03:29 crc kubenswrapper[4787]: E1001 10:03:29.502448 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="654cfd15-20ca-4754-9de7-ebe67905c8d2" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 01 10:03:29 crc kubenswrapper[4787]: I1001 10:03:29.502470 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="654cfd15-20ca-4754-9de7-ebe67905c8d2" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 01 10:03:29 crc kubenswrapper[4787]: I1001 10:03:29.502696 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="654cfd15-20ca-4754-9de7-ebe67905c8d2" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 01 10:03:29 crc kubenswrapper[4787]: I1001 10:03:29.503682 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zwc7x" Oct 01 10:03:29 crc kubenswrapper[4787]: I1001 10:03:29.506521 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vh5v6" Oct 01 10:03:29 crc kubenswrapper[4787]: I1001 10:03:29.506791 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 10:03:29 crc kubenswrapper[4787]: I1001 10:03:29.509790 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 10:03:29 crc kubenswrapper[4787]: I1001 10:03:29.510189 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 10:03:29 crc kubenswrapper[4787]: I1001 10:03:29.513936 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2b165b4-e2e9-4b0c-91f7-db75bcbc113d-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zwc7x\" (UID: \"c2b165b4-e2e9-4b0c-91f7-db75bcbc113d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zwc7x" Oct 01 10:03:29 crc kubenswrapper[4787]: I1001 10:03:29.520025 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j74hs\" (UniqueName: \"kubernetes.io/projected/c2b165b4-e2e9-4b0c-91f7-db75bcbc113d-kube-api-access-j74hs\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zwc7x\" (UID: \"c2b165b4-e2e9-4b0c-91f7-db75bcbc113d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zwc7x" Oct 01 10:03:29 crc kubenswrapper[4787]: I1001 10:03:29.517646 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zwc7x"] Oct 01 10:03:29 crc kubenswrapper[4787]: I1001 10:03:29.523840 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2b165b4-e2e9-4b0c-91f7-db75bcbc113d-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zwc7x\" (UID: \"c2b165b4-e2e9-4b0c-91f7-db75bcbc113d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zwc7x" Oct 01 10:03:29 crc kubenswrapper[4787]: I1001 10:03:29.625384 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2b165b4-e2e9-4b0c-91f7-db75bcbc113d-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zwc7x\" (UID: \"c2b165b4-e2e9-4b0c-91f7-db75bcbc113d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zwc7x" Oct 01 10:03:29 crc kubenswrapper[4787]: I1001 10:03:29.625516 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j74hs\" (UniqueName: \"kubernetes.io/projected/c2b165b4-e2e9-4b0c-91f7-db75bcbc113d-kube-api-access-j74hs\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zwc7x\" (UID: \"c2b165b4-e2e9-4b0c-91f7-db75bcbc113d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zwc7x" Oct 01 10:03:29 crc kubenswrapper[4787]: I1001 10:03:29.625608 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2b165b4-e2e9-4b0c-91f7-db75bcbc113d-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zwc7x\" (UID: \"c2b165b4-e2e9-4b0c-91f7-db75bcbc113d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zwc7x" Oct 01 10:03:29 crc kubenswrapper[4787]: I1001 10:03:29.630218 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2b165b4-e2e9-4b0c-91f7-db75bcbc113d-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zwc7x\" (UID: \"c2b165b4-e2e9-4b0c-91f7-db75bcbc113d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zwc7x" Oct 01 10:03:29 crc kubenswrapper[4787]: I1001 10:03:29.633824 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2b165b4-e2e9-4b0c-91f7-db75bcbc113d-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zwc7x\" (UID: \"c2b165b4-e2e9-4b0c-91f7-db75bcbc113d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zwc7x" Oct 01 10:03:29 crc kubenswrapper[4787]: I1001 10:03:29.642334 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j74hs\" (UniqueName: \"kubernetes.io/projected/c2b165b4-e2e9-4b0c-91f7-db75bcbc113d-kube-api-access-j74hs\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zwc7x\" (UID: \"c2b165b4-e2e9-4b0c-91f7-db75bcbc113d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zwc7x" Oct 01 10:03:29 crc kubenswrapper[4787]: I1001 10:03:29.823484 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zwc7x" Oct 01 10:03:30 crc kubenswrapper[4787]: I1001 10:03:30.327055 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zwc7x"] Oct 01 10:03:30 crc kubenswrapper[4787]: I1001 10:03:30.424802 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zwc7x" event={"ID":"c2b165b4-e2e9-4b0c-91f7-db75bcbc113d","Type":"ContainerStarted","Data":"a264e14526c3273a8265f641f1731bff58bb9d8025b98ae535fc6c6bde9b354d"} Oct 01 10:03:31 crc kubenswrapper[4787]: I1001 10:03:31.393451 4787 scope.go:117] "RemoveContainer" containerID="77f44314acff33e7ec38659c6b9db47146a92cd1eb9a9366d6be65d8a7e8d8ce" Oct 01 10:03:31 crc kubenswrapper[4787]: I1001 10:03:31.423199 4787 scope.go:117] "RemoveContainer" containerID="b9413fc51a4c3892f13d3a0c92848789cf1e05429822e02ff764a347dc393fdc" Oct 01 10:03:31 crc kubenswrapper[4787]: I1001 10:03:31.440624 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zwc7x" event={"ID":"c2b165b4-e2e9-4b0c-91f7-db75bcbc113d","Type":"ContainerStarted","Data":"1a1f5144cae27c7d55b95b4e47f807822124f83768ab1f590a0f5a65b587ba4c"} Oct 01 10:03:31 crc kubenswrapper[4787]: I1001 10:03:31.453767 4787 scope.go:117] "RemoveContainer" containerID="6c9d2abe0acc8461d7228c5f12c76da347f54d558bfe1892d9caa55030e4879a" Oct 01 10:03:31 crc kubenswrapper[4787]: I1001 10:03:31.469509 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zwc7x" podStartSLOduration=1.743965306 podStartE2EDuration="2.469482029s" podCreationTimestamp="2025-10-01 10:03:29 +0000 UTC" firstStartedPulling="2025-10-01 10:03:30.343971524 +0000 UTC m=+1642.459115681" lastFinishedPulling="2025-10-01 10:03:31.069488207 +0000 UTC m=+1643.184632404" observedRunningTime="2025-10-01 10:03:31.462336571 +0000 UTC m=+1643.577480748" watchObservedRunningTime="2025-10-01 10:03:31.469482029 +0000 UTC m=+1643.584626206" Oct 01 10:03:31 crc kubenswrapper[4787]: I1001 10:03:31.504413 4787 scope.go:117] "RemoveContainer" containerID="636e18451477b045320e69118c705fe6303e8717966fab13f8b4ae36f44898f5" Oct 01 10:03:31 crc kubenswrapper[4787]: I1001 10:03:31.577368 4787 scope.go:117] "RemoveContainer" containerID="0657daa702ef6496c3116fb6a7456ba0952ba05d4954c24ac71b5045b3e0c13b" Oct 01 10:03:31 crc kubenswrapper[4787]: I1001 10:03:31.628858 4787 scope.go:117] "RemoveContainer" containerID="f7622eb300e7d35468fcb5c5811f3de2c4ffb6c7e2df6c995cd08e0572b5abe1" Oct 01 10:03:31 crc kubenswrapper[4787]: I1001 10:03:31.671464 4787 scope.go:117] "RemoveContainer" containerID="998de087f5256977e33199f2b4e1481765a110036bd7b68cb2f9de0a37109a9e" Oct 01 10:03:32 crc kubenswrapper[4787]: I1001 10:03:32.026343 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-400d-account-create-mtvjl"] Oct 01 10:03:32 crc kubenswrapper[4787]: I1001 10:03:32.033639 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-400d-account-create-mtvjl"] Oct 01 10:03:32 crc kubenswrapper[4787]: I1001 10:03:32.524619 4787 scope.go:117] "RemoveContainer" containerID="e01333312cf21546b537b606d139d2d34e67fa8885c08f9ceb1f4c790ae4fd3a" Oct 01 10:03:32 crc kubenswrapper[4787]: E1001 10:03:32.524914 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:03:32 crc kubenswrapper[4787]: I1001 10:03:32.538691 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d768476-fd86-440c-9c5d-dd62b6cbcaa5" path="/var/lib/kubelet/pods/0d768476-fd86-440c-9c5d-dd62b6cbcaa5/volumes" Oct 01 10:03:42 crc kubenswrapper[4787]: I1001 10:03:42.037724 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-9f21-account-create-8zqzm"] Oct 01 10:03:42 crc kubenswrapper[4787]: I1001 10:03:42.055599 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-9192-account-create-rcrfw"] Oct 01 10:03:42 crc kubenswrapper[4787]: I1001 10:03:42.065608 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-9192-account-create-rcrfw"] Oct 01 10:03:42 crc kubenswrapper[4787]: I1001 10:03:42.074808 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-9f21-account-create-8zqzm"] Oct 01 10:03:42 crc kubenswrapper[4787]: I1001 10:03:42.534690 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00ce9c3d-e2af-4c96-a018-87539fbe9b42" path="/var/lib/kubelet/pods/00ce9c3d-e2af-4c96-a018-87539fbe9b42/volumes" Oct 01 10:03:42 crc kubenswrapper[4787]: I1001 10:03:42.535523 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77912320-f656-463d-a030-551ff526f530" path="/var/lib/kubelet/pods/77912320-f656-463d-a030-551ff526f530/volumes" Oct 01 10:03:47 crc kubenswrapper[4787]: I1001 10:03:47.524255 4787 scope.go:117] "RemoveContainer" containerID="e01333312cf21546b537b606d139d2d34e67fa8885c08f9ceb1f4c790ae4fd3a" Oct 01 10:03:47 crc kubenswrapper[4787]: E1001 10:03:47.524977 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:04:00 crc kubenswrapper[4787]: I1001 10:04:00.524010 4787 scope.go:117] "RemoveContainer" containerID="e01333312cf21546b537b606d139d2d34e67fa8885c08f9ceb1f4c790ae4fd3a" Oct 01 10:04:00 crc kubenswrapper[4787]: E1001 10:04:00.524734 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:04:05 crc kubenswrapper[4787]: I1001 10:04:05.052965 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-tj64b"] Oct 01 10:04:05 crc kubenswrapper[4787]: I1001 10:04:05.062775 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-tj64b"] Oct 01 10:04:06 crc kubenswrapper[4787]: I1001 10:04:06.536597 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9821f091-0513-4370-beb7-8239eab74972" path="/var/lib/kubelet/pods/9821f091-0513-4370-beb7-8239eab74972/volumes" Oct 01 10:04:15 crc kubenswrapper[4787]: I1001 10:04:15.524611 4787 scope.go:117] "RemoveContainer" containerID="e01333312cf21546b537b606d139d2d34e67fa8885c08f9ceb1f4c790ae4fd3a" Oct 01 10:04:15 crc kubenswrapper[4787]: E1001 10:04:15.525843 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:04:27 crc kubenswrapper[4787]: I1001 10:04:27.012003 4787 generic.go:334] "Generic (PLEG): container finished" podID="c2b165b4-e2e9-4b0c-91f7-db75bcbc113d" containerID="1a1f5144cae27c7d55b95b4e47f807822124f83768ab1f590a0f5a65b587ba4c" exitCode=2 Oct 01 10:04:27 crc kubenswrapper[4787]: I1001 10:04:27.012109 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zwc7x" event={"ID":"c2b165b4-e2e9-4b0c-91f7-db75bcbc113d","Type":"ContainerDied","Data":"1a1f5144cae27c7d55b95b4e47f807822124f83768ab1f590a0f5a65b587ba4c"} Oct 01 10:04:28 crc kubenswrapper[4787]: I1001 10:04:28.436519 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zwc7x" Oct 01 10:04:28 crc kubenswrapper[4787]: I1001 10:04:28.530545 4787 scope.go:117] "RemoveContainer" containerID="e01333312cf21546b537b606d139d2d34e67fa8885c08f9ceb1f4c790ae4fd3a" Oct 01 10:04:28 crc kubenswrapper[4787]: E1001 10:04:28.530987 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:04:28 crc kubenswrapper[4787]: I1001 10:04:28.583995 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2b165b4-e2e9-4b0c-91f7-db75bcbc113d-ssh-key\") pod \"c2b165b4-e2e9-4b0c-91f7-db75bcbc113d\" (UID: \"c2b165b4-e2e9-4b0c-91f7-db75bcbc113d\") " Oct 01 10:04:28 crc kubenswrapper[4787]: I1001 10:04:28.584061 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2b165b4-e2e9-4b0c-91f7-db75bcbc113d-inventory\") pod \"c2b165b4-e2e9-4b0c-91f7-db75bcbc113d\" (UID: \"c2b165b4-e2e9-4b0c-91f7-db75bcbc113d\") " Oct 01 10:04:28 crc kubenswrapper[4787]: I1001 10:04:28.584268 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j74hs\" (UniqueName: \"kubernetes.io/projected/c2b165b4-e2e9-4b0c-91f7-db75bcbc113d-kube-api-access-j74hs\") pod \"c2b165b4-e2e9-4b0c-91f7-db75bcbc113d\" (UID: \"c2b165b4-e2e9-4b0c-91f7-db75bcbc113d\") " Oct 01 10:04:28 crc kubenswrapper[4787]: I1001 10:04:28.593671 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2b165b4-e2e9-4b0c-91f7-db75bcbc113d-kube-api-access-j74hs" (OuterVolumeSpecName: "kube-api-access-j74hs") pod "c2b165b4-e2e9-4b0c-91f7-db75bcbc113d" (UID: "c2b165b4-e2e9-4b0c-91f7-db75bcbc113d"). InnerVolumeSpecName "kube-api-access-j74hs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:04:28 crc kubenswrapper[4787]: I1001 10:04:28.614737 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2b165b4-e2e9-4b0c-91f7-db75bcbc113d-inventory" (OuterVolumeSpecName: "inventory") pod "c2b165b4-e2e9-4b0c-91f7-db75bcbc113d" (UID: "c2b165b4-e2e9-4b0c-91f7-db75bcbc113d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:04:28 crc kubenswrapper[4787]: I1001 10:04:28.633666 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2b165b4-e2e9-4b0c-91f7-db75bcbc113d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c2b165b4-e2e9-4b0c-91f7-db75bcbc113d" (UID: "c2b165b4-e2e9-4b0c-91f7-db75bcbc113d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:04:28 crc kubenswrapper[4787]: I1001 10:04:28.689320 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j74hs\" (UniqueName: \"kubernetes.io/projected/c2b165b4-e2e9-4b0c-91f7-db75bcbc113d-kube-api-access-j74hs\") on node \"crc\" DevicePath \"\"" Oct 01 10:04:28 crc kubenswrapper[4787]: I1001 10:04:28.689372 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2b165b4-e2e9-4b0c-91f7-db75bcbc113d-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 10:04:28 crc kubenswrapper[4787]: I1001 10:04:28.689385 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2b165b4-e2e9-4b0c-91f7-db75bcbc113d-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 10:04:29 crc kubenswrapper[4787]: I1001 10:04:29.032933 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zwc7x" event={"ID":"c2b165b4-e2e9-4b0c-91f7-db75bcbc113d","Type":"ContainerDied","Data":"a264e14526c3273a8265f641f1731bff58bb9d8025b98ae535fc6c6bde9b354d"} Oct 01 10:04:29 crc kubenswrapper[4787]: I1001 10:04:29.032981 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a264e14526c3273a8265f641f1731bff58bb9d8025b98ae535fc6c6bde9b354d" Oct 01 10:04:29 crc kubenswrapper[4787]: I1001 10:04:29.033023 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zwc7x" Oct 01 10:04:29 crc kubenswrapper[4787]: I1001 10:04:29.061570 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-xglhv"] Oct 01 10:04:29 crc kubenswrapper[4787]: I1001 10:04:29.081062 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-jqjq6"] Oct 01 10:04:29 crc kubenswrapper[4787]: I1001 10:04:29.091234 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-xglhv"] Oct 01 10:04:29 crc kubenswrapper[4787]: I1001 10:04:29.097413 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-jqjq6"] Oct 01 10:04:30 crc kubenswrapper[4787]: I1001 10:04:30.534703 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0634a251-9095-454e-8a37-ff57c2ec40e1" path="/var/lib/kubelet/pods/0634a251-9095-454e-8a37-ff57c2ec40e1/volumes" Oct 01 10:04:30 crc kubenswrapper[4787]: I1001 10:04:30.535892 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6ac40b1-e2a4-4d15-9855-6bbf9daad643" path="/var/lib/kubelet/pods/a6ac40b1-e2a4-4d15-9855-6bbf9daad643/volumes" Oct 01 10:04:31 crc kubenswrapper[4787]: I1001 10:04:31.882415 4787 scope.go:117] "RemoveContainer" containerID="f1b49a0aadf27cc2a040d9304a795968e93c53ef38c082785703c4bf0b203b31" Oct 01 10:04:31 crc kubenswrapper[4787]: I1001 10:04:31.936044 4787 scope.go:117] "RemoveContainer" containerID="a4ea6073b367bb6bb5a6b790a4f7e3b223a52f0e0035eef4de083e15ebddf3b8" Oct 01 10:04:31 crc kubenswrapper[4787]: I1001 10:04:31.966874 4787 scope.go:117] "RemoveContainer" containerID="431a329239e735143b5e81c9f3853dc3171c5a42ceda423213bbd0424a81afa7" Oct 01 10:04:32 crc kubenswrapper[4787]: I1001 10:04:32.009037 4787 scope.go:117] "RemoveContainer" containerID="515b5403724fa4d69f5a7ff8da7ef602083269aeb1885f1f9803483e65cb522d" Oct 01 10:04:32 crc kubenswrapper[4787]: I1001 10:04:32.075804 4787 scope.go:117] "RemoveContainer" containerID="669a8eed122a993a2d627ab48625d872ec440697cf9206734d572865c06fccbf" Oct 01 10:04:32 crc kubenswrapper[4787]: I1001 10:04:32.130513 4787 scope.go:117] "RemoveContainer" containerID="39c83d707b3a7ad213890cc04ebe3e5334895c38861f85f05c13ed7f5a73cb19" Oct 01 10:04:36 crc kubenswrapper[4787]: I1001 10:04:36.037099 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddl72"] Oct 01 10:04:36 crc kubenswrapper[4787]: E1001 10:04:36.038185 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2b165b4-e2e9-4b0c-91f7-db75bcbc113d" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 01 10:04:36 crc kubenswrapper[4787]: I1001 10:04:36.038204 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2b165b4-e2e9-4b0c-91f7-db75bcbc113d" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 01 10:04:36 crc kubenswrapper[4787]: I1001 10:04:36.038442 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2b165b4-e2e9-4b0c-91f7-db75bcbc113d" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 01 10:04:36 crc kubenswrapper[4787]: I1001 10:04:36.039269 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddl72" Oct 01 10:04:36 crc kubenswrapper[4787]: I1001 10:04:36.044198 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 10:04:36 crc kubenswrapper[4787]: I1001 10:04:36.044507 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 10:04:36 crc kubenswrapper[4787]: I1001 10:04:36.045662 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vh5v6" Oct 01 10:04:36 crc kubenswrapper[4787]: I1001 10:04:36.048681 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddl72"] Oct 01 10:04:36 crc kubenswrapper[4787]: I1001 10:04:36.049541 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 10:04:36 crc kubenswrapper[4787]: I1001 10:04:36.135148 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/def7d40c-8191-4256-b8ad-4fe9272d5fae-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ddl72\" (UID: \"def7d40c-8191-4256-b8ad-4fe9272d5fae\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddl72" Oct 01 10:04:36 crc kubenswrapper[4787]: I1001 10:04:36.135241 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmjtc\" (UniqueName: \"kubernetes.io/projected/def7d40c-8191-4256-b8ad-4fe9272d5fae-kube-api-access-lmjtc\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ddl72\" (UID: \"def7d40c-8191-4256-b8ad-4fe9272d5fae\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddl72" Oct 01 10:04:36 crc kubenswrapper[4787]: I1001 10:04:36.135284 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/def7d40c-8191-4256-b8ad-4fe9272d5fae-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ddl72\" (UID: \"def7d40c-8191-4256-b8ad-4fe9272d5fae\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddl72" Oct 01 10:04:36 crc kubenswrapper[4787]: I1001 10:04:36.237558 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/def7d40c-8191-4256-b8ad-4fe9272d5fae-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ddl72\" (UID: \"def7d40c-8191-4256-b8ad-4fe9272d5fae\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddl72" Oct 01 10:04:36 crc kubenswrapper[4787]: I1001 10:04:36.237645 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmjtc\" (UniqueName: \"kubernetes.io/projected/def7d40c-8191-4256-b8ad-4fe9272d5fae-kube-api-access-lmjtc\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ddl72\" (UID: \"def7d40c-8191-4256-b8ad-4fe9272d5fae\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddl72" Oct 01 10:04:36 crc kubenswrapper[4787]: I1001 10:04:36.237709 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/def7d40c-8191-4256-b8ad-4fe9272d5fae-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ddl72\" (UID: \"def7d40c-8191-4256-b8ad-4fe9272d5fae\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddl72" Oct 01 10:04:36 crc kubenswrapper[4787]: I1001 10:04:36.244876 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/def7d40c-8191-4256-b8ad-4fe9272d5fae-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ddl72\" (UID: \"def7d40c-8191-4256-b8ad-4fe9272d5fae\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddl72" Oct 01 10:04:36 crc kubenswrapper[4787]: I1001 10:04:36.253064 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/def7d40c-8191-4256-b8ad-4fe9272d5fae-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ddl72\" (UID: \"def7d40c-8191-4256-b8ad-4fe9272d5fae\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddl72" Oct 01 10:04:36 crc kubenswrapper[4787]: I1001 10:04:36.253916 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmjtc\" (UniqueName: \"kubernetes.io/projected/def7d40c-8191-4256-b8ad-4fe9272d5fae-kube-api-access-lmjtc\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ddl72\" (UID: \"def7d40c-8191-4256-b8ad-4fe9272d5fae\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddl72" Oct 01 10:04:36 crc kubenswrapper[4787]: I1001 10:04:36.394171 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddl72" Oct 01 10:04:36 crc kubenswrapper[4787]: I1001 10:04:36.942315 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddl72"] Oct 01 10:04:37 crc kubenswrapper[4787]: I1001 10:04:37.116007 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddl72" event={"ID":"def7d40c-8191-4256-b8ad-4fe9272d5fae","Type":"ContainerStarted","Data":"bb88410da72811e7761ec26bd9f843293c0fd67a18790c8431ee74cf04f5b34a"} Oct 01 10:04:38 crc kubenswrapper[4787]: I1001 10:04:38.124771 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddl72" event={"ID":"def7d40c-8191-4256-b8ad-4fe9272d5fae","Type":"ContainerStarted","Data":"b3316afd1726ee0b819fc158107edf857a0ccad52fa77f8f2837a0f9a0cd5992"} Oct 01 10:04:38 crc kubenswrapper[4787]: I1001 10:04:38.163564 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddl72" podStartSLOduration=1.662736862 podStartE2EDuration="2.163541173s" podCreationTimestamp="2025-10-01 10:04:36 +0000 UTC" firstStartedPulling="2025-10-01 10:04:36.94847104 +0000 UTC m=+1709.063615197" lastFinishedPulling="2025-10-01 10:04:37.449275351 +0000 UTC m=+1709.564419508" observedRunningTime="2025-10-01 10:04:38.156440486 +0000 UTC m=+1710.271584683" watchObservedRunningTime="2025-10-01 10:04:38.163541173 +0000 UTC m=+1710.278685330" Oct 01 10:04:39 crc kubenswrapper[4787]: I1001 10:04:39.524127 4787 scope.go:117] "RemoveContainer" containerID="e01333312cf21546b537b606d139d2d34e67fa8885c08f9ceb1f4c790ae4fd3a" Oct 01 10:04:39 crc kubenswrapper[4787]: E1001 10:04:39.524652 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:04:50 crc kubenswrapper[4787]: I1001 10:04:50.524403 4787 scope.go:117] "RemoveContainer" containerID="e01333312cf21546b537b606d139d2d34e67fa8885c08f9ceb1f4c790ae4fd3a" Oct 01 10:04:50 crc kubenswrapper[4787]: E1001 10:04:50.525259 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:05:04 crc kubenswrapper[4787]: I1001 10:05:04.523999 4787 scope.go:117] "RemoveContainer" containerID="e01333312cf21546b537b606d139d2d34e67fa8885c08f9ceb1f4c790ae4fd3a" Oct 01 10:05:04 crc kubenswrapper[4787]: E1001 10:05:04.524868 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:05:13 crc kubenswrapper[4787]: I1001 10:05:13.043517 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-8655l"] Oct 01 10:05:13 crc kubenswrapper[4787]: I1001 10:05:13.050166 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-8655l"] Oct 01 10:05:14 crc kubenswrapper[4787]: I1001 10:05:14.534895 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5d060f6-9fac-4c41-b365-033b09e68016" path="/var/lib/kubelet/pods/f5d060f6-9fac-4c41-b365-033b09e68016/volumes" Oct 01 10:05:19 crc kubenswrapper[4787]: I1001 10:05:19.524208 4787 scope.go:117] "RemoveContainer" containerID="e01333312cf21546b537b606d139d2d34e67fa8885c08f9ceb1f4c790ae4fd3a" Oct 01 10:05:19 crc kubenswrapper[4787]: E1001 10:05:19.525828 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:05:21 crc kubenswrapper[4787]: I1001 10:05:21.501360 4787 generic.go:334] "Generic (PLEG): container finished" podID="def7d40c-8191-4256-b8ad-4fe9272d5fae" containerID="b3316afd1726ee0b819fc158107edf857a0ccad52fa77f8f2837a0f9a0cd5992" exitCode=0 Oct 01 10:05:21 crc kubenswrapper[4787]: I1001 10:05:21.501487 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddl72" event={"ID":"def7d40c-8191-4256-b8ad-4fe9272d5fae","Type":"ContainerDied","Data":"b3316afd1726ee0b819fc158107edf857a0ccad52fa77f8f2837a0f9a0cd5992"} Oct 01 10:05:22 crc kubenswrapper[4787]: I1001 10:05:22.861928 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddl72" Oct 01 10:05:23 crc kubenswrapper[4787]: I1001 10:05:23.001840 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/def7d40c-8191-4256-b8ad-4fe9272d5fae-ssh-key\") pod \"def7d40c-8191-4256-b8ad-4fe9272d5fae\" (UID: \"def7d40c-8191-4256-b8ad-4fe9272d5fae\") " Oct 01 10:05:23 crc kubenswrapper[4787]: I1001 10:05:23.001933 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/def7d40c-8191-4256-b8ad-4fe9272d5fae-inventory\") pod \"def7d40c-8191-4256-b8ad-4fe9272d5fae\" (UID: \"def7d40c-8191-4256-b8ad-4fe9272d5fae\") " Oct 01 10:05:23 crc kubenswrapper[4787]: I1001 10:05:23.001993 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmjtc\" (UniqueName: \"kubernetes.io/projected/def7d40c-8191-4256-b8ad-4fe9272d5fae-kube-api-access-lmjtc\") pod \"def7d40c-8191-4256-b8ad-4fe9272d5fae\" (UID: \"def7d40c-8191-4256-b8ad-4fe9272d5fae\") " Oct 01 10:05:23 crc kubenswrapper[4787]: I1001 10:05:23.008300 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/def7d40c-8191-4256-b8ad-4fe9272d5fae-kube-api-access-lmjtc" (OuterVolumeSpecName: "kube-api-access-lmjtc") pod "def7d40c-8191-4256-b8ad-4fe9272d5fae" (UID: "def7d40c-8191-4256-b8ad-4fe9272d5fae"). InnerVolumeSpecName "kube-api-access-lmjtc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:05:23 crc kubenswrapper[4787]: I1001 10:05:23.028729 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/def7d40c-8191-4256-b8ad-4fe9272d5fae-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "def7d40c-8191-4256-b8ad-4fe9272d5fae" (UID: "def7d40c-8191-4256-b8ad-4fe9272d5fae"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:05:23 crc kubenswrapper[4787]: I1001 10:05:23.041448 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/def7d40c-8191-4256-b8ad-4fe9272d5fae-inventory" (OuterVolumeSpecName: "inventory") pod "def7d40c-8191-4256-b8ad-4fe9272d5fae" (UID: "def7d40c-8191-4256-b8ad-4fe9272d5fae"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:05:23 crc kubenswrapper[4787]: I1001 10:05:23.104954 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/def7d40c-8191-4256-b8ad-4fe9272d5fae-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:23 crc kubenswrapper[4787]: I1001 10:05:23.104996 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/def7d40c-8191-4256-b8ad-4fe9272d5fae-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:23 crc kubenswrapper[4787]: I1001 10:05:23.105010 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmjtc\" (UniqueName: \"kubernetes.io/projected/def7d40c-8191-4256-b8ad-4fe9272d5fae-kube-api-access-lmjtc\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:23 crc kubenswrapper[4787]: I1001 10:05:23.517085 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddl72" event={"ID":"def7d40c-8191-4256-b8ad-4fe9272d5fae","Type":"ContainerDied","Data":"bb88410da72811e7761ec26bd9f843293c0fd67a18790c8431ee74cf04f5b34a"} Oct 01 10:05:23 crc kubenswrapper[4787]: I1001 10:05:23.518523 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb88410da72811e7761ec26bd9f843293c0fd67a18790c8431ee74cf04f5b34a" Oct 01 10:05:23 crc kubenswrapper[4787]: I1001 10:05:23.517120 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ddl72" Oct 01 10:05:23 crc kubenswrapper[4787]: I1001 10:05:23.608977 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-rhhxm"] Oct 01 10:05:23 crc kubenswrapper[4787]: E1001 10:05:23.609477 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="def7d40c-8191-4256-b8ad-4fe9272d5fae" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 01 10:05:23 crc kubenswrapper[4787]: I1001 10:05:23.609508 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="def7d40c-8191-4256-b8ad-4fe9272d5fae" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 01 10:05:23 crc kubenswrapper[4787]: I1001 10:05:23.609801 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="def7d40c-8191-4256-b8ad-4fe9272d5fae" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 01 10:05:23 crc kubenswrapper[4787]: I1001 10:05:23.610683 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-rhhxm" Oct 01 10:05:23 crc kubenswrapper[4787]: I1001 10:05:23.616668 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 10:05:23 crc kubenswrapper[4787]: I1001 10:05:23.618532 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 10:05:23 crc kubenswrapper[4787]: I1001 10:05:23.619492 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vh5v6" Oct 01 10:05:23 crc kubenswrapper[4787]: I1001 10:05:23.623438 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 10:05:23 crc kubenswrapper[4787]: I1001 10:05:23.637760 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-rhhxm"] Oct 01 10:05:23 crc kubenswrapper[4787]: E1001 10:05:23.715214 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddef7d40c_8191_4256_b8ad_4fe9272d5fae.slice/crio-bb88410da72811e7761ec26bd9f843293c0fd67a18790c8431ee74cf04f5b34a\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddef7d40c_8191_4256_b8ad_4fe9272d5fae.slice\": RecentStats: unable to find data in memory cache]" Oct 01 10:05:23 crc kubenswrapper[4787]: I1001 10:05:23.740737 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/49e3ae1d-7abd-4e22-a333-54e10db349a6-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-rhhxm\" (UID: \"49e3ae1d-7abd-4e22-a333-54e10db349a6\") " pod="openstack/ssh-known-hosts-edpm-deployment-rhhxm" Oct 01 10:05:23 crc kubenswrapper[4787]: I1001 10:05:23.740882 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zpl4\" (UniqueName: \"kubernetes.io/projected/49e3ae1d-7abd-4e22-a333-54e10db349a6-kube-api-access-4zpl4\") pod \"ssh-known-hosts-edpm-deployment-rhhxm\" (UID: \"49e3ae1d-7abd-4e22-a333-54e10db349a6\") " pod="openstack/ssh-known-hosts-edpm-deployment-rhhxm" Oct 01 10:05:23 crc kubenswrapper[4787]: I1001 10:05:23.740916 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/49e3ae1d-7abd-4e22-a333-54e10db349a6-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-rhhxm\" (UID: \"49e3ae1d-7abd-4e22-a333-54e10db349a6\") " pod="openstack/ssh-known-hosts-edpm-deployment-rhhxm" Oct 01 10:05:23 crc kubenswrapper[4787]: I1001 10:05:23.846156 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zpl4\" (UniqueName: \"kubernetes.io/projected/49e3ae1d-7abd-4e22-a333-54e10db349a6-kube-api-access-4zpl4\") pod \"ssh-known-hosts-edpm-deployment-rhhxm\" (UID: \"49e3ae1d-7abd-4e22-a333-54e10db349a6\") " pod="openstack/ssh-known-hosts-edpm-deployment-rhhxm" Oct 01 10:05:23 crc kubenswrapper[4787]: I1001 10:05:23.846438 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/49e3ae1d-7abd-4e22-a333-54e10db349a6-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-rhhxm\" (UID: \"49e3ae1d-7abd-4e22-a333-54e10db349a6\") " pod="openstack/ssh-known-hosts-edpm-deployment-rhhxm" Oct 01 10:05:23 crc kubenswrapper[4787]: I1001 10:05:23.846572 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/49e3ae1d-7abd-4e22-a333-54e10db349a6-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-rhhxm\" (UID: \"49e3ae1d-7abd-4e22-a333-54e10db349a6\") " pod="openstack/ssh-known-hosts-edpm-deployment-rhhxm" Oct 01 10:05:23 crc kubenswrapper[4787]: I1001 10:05:23.851201 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/49e3ae1d-7abd-4e22-a333-54e10db349a6-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-rhhxm\" (UID: \"49e3ae1d-7abd-4e22-a333-54e10db349a6\") " pod="openstack/ssh-known-hosts-edpm-deployment-rhhxm" Oct 01 10:05:23 crc kubenswrapper[4787]: I1001 10:05:23.855560 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/49e3ae1d-7abd-4e22-a333-54e10db349a6-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-rhhxm\" (UID: \"49e3ae1d-7abd-4e22-a333-54e10db349a6\") " pod="openstack/ssh-known-hosts-edpm-deployment-rhhxm" Oct 01 10:05:23 crc kubenswrapper[4787]: I1001 10:05:23.864756 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zpl4\" (UniqueName: \"kubernetes.io/projected/49e3ae1d-7abd-4e22-a333-54e10db349a6-kube-api-access-4zpl4\") pod \"ssh-known-hosts-edpm-deployment-rhhxm\" (UID: \"49e3ae1d-7abd-4e22-a333-54e10db349a6\") " pod="openstack/ssh-known-hosts-edpm-deployment-rhhxm" Oct 01 10:05:23 crc kubenswrapper[4787]: I1001 10:05:23.938779 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-rhhxm" Oct 01 10:05:24 crc kubenswrapper[4787]: I1001 10:05:24.485874 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-rhhxm"] Oct 01 10:05:24 crc kubenswrapper[4787]: I1001 10:05:24.534745 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-rhhxm" event={"ID":"49e3ae1d-7abd-4e22-a333-54e10db349a6","Type":"ContainerStarted","Data":"ed61d8dc63c51624b26fd76b7681c5050731565852012e5849b9053f54d89423"} Oct 01 10:05:25 crc kubenswrapper[4787]: I1001 10:05:25.537401 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-rhhxm" event={"ID":"49e3ae1d-7abd-4e22-a333-54e10db349a6","Type":"ContainerStarted","Data":"0b49be10ba0056f14e37a608f492c3d04bc5cea5f0ca5ebd6fcd71ed41bbe759"} Oct 01 10:05:25 crc kubenswrapper[4787]: I1001 10:05:25.556379 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-rhhxm" podStartSLOduration=1.980862928 podStartE2EDuration="2.556360297s" podCreationTimestamp="2025-10-01 10:05:23 +0000 UTC" firstStartedPulling="2025-10-01 10:05:24.490528968 +0000 UTC m=+1756.605673155" lastFinishedPulling="2025-10-01 10:05:25.066026367 +0000 UTC m=+1757.181170524" observedRunningTime="2025-10-01 10:05:25.554974693 +0000 UTC m=+1757.670118870" watchObservedRunningTime="2025-10-01 10:05:25.556360297 +0000 UTC m=+1757.671504454" Oct 01 10:05:31 crc kubenswrapper[4787]: I1001 10:05:31.588547 4787 generic.go:334] "Generic (PLEG): container finished" podID="49e3ae1d-7abd-4e22-a333-54e10db349a6" containerID="0b49be10ba0056f14e37a608f492c3d04bc5cea5f0ca5ebd6fcd71ed41bbe759" exitCode=0 Oct 01 10:05:31 crc kubenswrapper[4787]: I1001 10:05:31.588631 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-rhhxm" event={"ID":"49e3ae1d-7abd-4e22-a333-54e10db349a6","Type":"ContainerDied","Data":"0b49be10ba0056f14e37a608f492c3d04bc5cea5f0ca5ebd6fcd71ed41bbe759"} Oct 01 10:05:32 crc kubenswrapper[4787]: I1001 10:05:32.264959 4787 scope.go:117] "RemoveContainer" containerID="9d52221a28c0b16ca76146635754207f0869a14d5e5187ba7d4afc4661e743a4" Oct 01 10:05:32 crc kubenswrapper[4787]: I1001 10:05:32.524616 4787 scope.go:117] "RemoveContainer" containerID="e01333312cf21546b537b606d139d2d34e67fa8885c08f9ceb1f4c790ae4fd3a" Oct 01 10:05:32 crc kubenswrapper[4787]: E1001 10:05:32.524875 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:05:33 crc kubenswrapper[4787]: I1001 10:05:33.056193 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-rhhxm" Oct 01 10:05:33 crc kubenswrapper[4787]: I1001 10:05:33.216176 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/49e3ae1d-7abd-4e22-a333-54e10db349a6-ssh-key-openstack-edpm-ipam\") pod \"49e3ae1d-7abd-4e22-a333-54e10db349a6\" (UID: \"49e3ae1d-7abd-4e22-a333-54e10db349a6\") " Oct 01 10:05:33 crc kubenswrapper[4787]: I1001 10:05:33.216289 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zpl4\" (UniqueName: \"kubernetes.io/projected/49e3ae1d-7abd-4e22-a333-54e10db349a6-kube-api-access-4zpl4\") pod \"49e3ae1d-7abd-4e22-a333-54e10db349a6\" (UID: \"49e3ae1d-7abd-4e22-a333-54e10db349a6\") " Oct 01 10:05:33 crc kubenswrapper[4787]: I1001 10:05:33.216354 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/49e3ae1d-7abd-4e22-a333-54e10db349a6-inventory-0\") pod \"49e3ae1d-7abd-4e22-a333-54e10db349a6\" (UID: \"49e3ae1d-7abd-4e22-a333-54e10db349a6\") " Oct 01 10:05:33 crc kubenswrapper[4787]: I1001 10:05:33.224533 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49e3ae1d-7abd-4e22-a333-54e10db349a6-kube-api-access-4zpl4" (OuterVolumeSpecName: "kube-api-access-4zpl4") pod "49e3ae1d-7abd-4e22-a333-54e10db349a6" (UID: "49e3ae1d-7abd-4e22-a333-54e10db349a6"). InnerVolumeSpecName "kube-api-access-4zpl4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:05:33 crc kubenswrapper[4787]: I1001 10:05:33.247094 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49e3ae1d-7abd-4e22-a333-54e10db349a6-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "49e3ae1d-7abd-4e22-a333-54e10db349a6" (UID: "49e3ae1d-7abd-4e22-a333-54e10db349a6"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:05:33 crc kubenswrapper[4787]: I1001 10:05:33.254051 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49e3ae1d-7abd-4e22-a333-54e10db349a6-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "49e3ae1d-7abd-4e22-a333-54e10db349a6" (UID: "49e3ae1d-7abd-4e22-a333-54e10db349a6"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:05:33 crc kubenswrapper[4787]: I1001 10:05:33.321623 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/49e3ae1d-7abd-4e22-a333-54e10db349a6-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:33 crc kubenswrapper[4787]: I1001 10:05:33.321692 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zpl4\" (UniqueName: \"kubernetes.io/projected/49e3ae1d-7abd-4e22-a333-54e10db349a6-kube-api-access-4zpl4\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:33 crc kubenswrapper[4787]: I1001 10:05:33.321727 4787 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/49e3ae1d-7abd-4e22-a333-54e10db349a6-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:33 crc kubenswrapper[4787]: I1001 10:05:33.607815 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-rhhxm" event={"ID":"49e3ae1d-7abd-4e22-a333-54e10db349a6","Type":"ContainerDied","Data":"ed61d8dc63c51624b26fd76b7681c5050731565852012e5849b9053f54d89423"} Oct 01 10:05:33 crc kubenswrapper[4787]: I1001 10:05:33.608119 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed61d8dc63c51624b26fd76b7681c5050731565852012e5849b9053f54d89423" Oct 01 10:05:33 crc kubenswrapper[4787]: I1001 10:05:33.607848 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-rhhxm" Oct 01 10:05:33 crc kubenswrapper[4787]: I1001 10:05:33.682584 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-ws2cx"] Oct 01 10:05:33 crc kubenswrapper[4787]: E1001 10:05:33.683205 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49e3ae1d-7abd-4e22-a333-54e10db349a6" containerName="ssh-known-hosts-edpm-deployment" Oct 01 10:05:33 crc kubenswrapper[4787]: I1001 10:05:33.683228 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="49e3ae1d-7abd-4e22-a333-54e10db349a6" containerName="ssh-known-hosts-edpm-deployment" Oct 01 10:05:33 crc kubenswrapper[4787]: I1001 10:05:33.683455 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="49e3ae1d-7abd-4e22-a333-54e10db349a6" containerName="ssh-known-hosts-edpm-deployment" Oct 01 10:05:33 crc kubenswrapper[4787]: I1001 10:05:33.684251 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ws2cx" Oct 01 10:05:33 crc kubenswrapper[4787]: I1001 10:05:33.687342 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 10:05:33 crc kubenswrapper[4787]: I1001 10:05:33.687392 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 10:05:33 crc kubenswrapper[4787]: I1001 10:05:33.692225 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vh5v6" Oct 01 10:05:33 crc kubenswrapper[4787]: I1001 10:05:33.692278 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 10:05:33 crc kubenswrapper[4787]: I1001 10:05:33.694873 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-ws2cx"] Oct 01 10:05:33 crc kubenswrapper[4787]: I1001 10:05:33.830791 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/389d6832-2d3e-49a0-afbd-f88359db6324-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ws2cx\" (UID: \"389d6832-2d3e-49a0-afbd-f88359db6324\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ws2cx" Oct 01 10:05:33 crc kubenswrapper[4787]: I1001 10:05:33.830893 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/389d6832-2d3e-49a0-afbd-f88359db6324-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ws2cx\" (UID: \"389d6832-2d3e-49a0-afbd-f88359db6324\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ws2cx" Oct 01 10:05:33 crc kubenswrapper[4787]: I1001 10:05:33.831160 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvk6s\" (UniqueName: \"kubernetes.io/projected/389d6832-2d3e-49a0-afbd-f88359db6324-kube-api-access-bvk6s\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ws2cx\" (UID: \"389d6832-2d3e-49a0-afbd-f88359db6324\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ws2cx" Oct 01 10:05:33 crc kubenswrapper[4787]: I1001 10:05:33.932464 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/389d6832-2d3e-49a0-afbd-f88359db6324-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ws2cx\" (UID: \"389d6832-2d3e-49a0-afbd-f88359db6324\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ws2cx" Oct 01 10:05:33 crc kubenswrapper[4787]: I1001 10:05:33.932577 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/389d6832-2d3e-49a0-afbd-f88359db6324-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ws2cx\" (UID: \"389d6832-2d3e-49a0-afbd-f88359db6324\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ws2cx" Oct 01 10:05:33 crc kubenswrapper[4787]: I1001 10:05:33.932614 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvk6s\" (UniqueName: \"kubernetes.io/projected/389d6832-2d3e-49a0-afbd-f88359db6324-kube-api-access-bvk6s\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ws2cx\" (UID: \"389d6832-2d3e-49a0-afbd-f88359db6324\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ws2cx" Oct 01 10:05:33 crc kubenswrapper[4787]: I1001 10:05:33.937920 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/389d6832-2d3e-49a0-afbd-f88359db6324-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ws2cx\" (UID: \"389d6832-2d3e-49a0-afbd-f88359db6324\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ws2cx" Oct 01 10:05:33 crc kubenswrapper[4787]: I1001 10:05:33.949722 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/389d6832-2d3e-49a0-afbd-f88359db6324-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ws2cx\" (UID: \"389d6832-2d3e-49a0-afbd-f88359db6324\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ws2cx" Oct 01 10:05:33 crc kubenswrapper[4787]: I1001 10:05:33.962900 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvk6s\" (UniqueName: \"kubernetes.io/projected/389d6832-2d3e-49a0-afbd-f88359db6324-kube-api-access-bvk6s\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ws2cx\" (UID: \"389d6832-2d3e-49a0-afbd-f88359db6324\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ws2cx" Oct 01 10:05:34 crc kubenswrapper[4787]: I1001 10:05:34.000927 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ws2cx" Oct 01 10:05:34 crc kubenswrapper[4787]: I1001 10:05:34.516131 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-ws2cx"] Oct 01 10:05:34 crc kubenswrapper[4787]: I1001 10:05:34.618156 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ws2cx" event={"ID":"389d6832-2d3e-49a0-afbd-f88359db6324","Type":"ContainerStarted","Data":"4388f421744e442f6afd5c31d56eac8753241bbae9b25176a6f7582999009354"} Oct 01 10:05:35 crc kubenswrapper[4787]: I1001 10:05:35.627602 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ws2cx" event={"ID":"389d6832-2d3e-49a0-afbd-f88359db6324","Type":"ContainerStarted","Data":"942c80097be1671d7bf68cc7e846c88c0167f35e39a6f79f60026651079beeea"} Oct 01 10:05:35 crc kubenswrapper[4787]: I1001 10:05:35.646270 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ws2cx" podStartSLOduration=2.076000413 podStartE2EDuration="2.646250293s" podCreationTimestamp="2025-10-01 10:05:33 +0000 UTC" firstStartedPulling="2025-10-01 10:05:34.526382128 +0000 UTC m=+1766.641526285" lastFinishedPulling="2025-10-01 10:05:35.096631998 +0000 UTC m=+1767.211776165" observedRunningTime="2025-10-01 10:05:35.64332524 +0000 UTC m=+1767.758469407" watchObservedRunningTime="2025-10-01 10:05:35.646250293 +0000 UTC m=+1767.761394450" Oct 01 10:05:43 crc kubenswrapper[4787]: I1001 10:05:43.694149 4787 generic.go:334] "Generic (PLEG): container finished" podID="389d6832-2d3e-49a0-afbd-f88359db6324" containerID="942c80097be1671d7bf68cc7e846c88c0167f35e39a6f79f60026651079beeea" exitCode=0 Oct 01 10:05:43 crc kubenswrapper[4787]: I1001 10:05:43.694270 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ws2cx" event={"ID":"389d6832-2d3e-49a0-afbd-f88359db6324","Type":"ContainerDied","Data":"942c80097be1671d7bf68cc7e846c88c0167f35e39a6f79f60026651079beeea"} Oct 01 10:05:45 crc kubenswrapper[4787]: I1001 10:05:45.134869 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ws2cx" Oct 01 10:05:45 crc kubenswrapper[4787]: I1001 10:05:45.145735 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/389d6832-2d3e-49a0-afbd-f88359db6324-inventory\") pod \"389d6832-2d3e-49a0-afbd-f88359db6324\" (UID: \"389d6832-2d3e-49a0-afbd-f88359db6324\") " Oct 01 10:05:45 crc kubenswrapper[4787]: I1001 10:05:45.146012 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvk6s\" (UniqueName: \"kubernetes.io/projected/389d6832-2d3e-49a0-afbd-f88359db6324-kube-api-access-bvk6s\") pod \"389d6832-2d3e-49a0-afbd-f88359db6324\" (UID: \"389d6832-2d3e-49a0-afbd-f88359db6324\") " Oct 01 10:05:45 crc kubenswrapper[4787]: I1001 10:05:45.147471 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/389d6832-2d3e-49a0-afbd-f88359db6324-ssh-key\") pod \"389d6832-2d3e-49a0-afbd-f88359db6324\" (UID: \"389d6832-2d3e-49a0-afbd-f88359db6324\") " Oct 01 10:05:45 crc kubenswrapper[4787]: I1001 10:05:45.159179 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/389d6832-2d3e-49a0-afbd-f88359db6324-kube-api-access-bvk6s" (OuterVolumeSpecName: "kube-api-access-bvk6s") pod "389d6832-2d3e-49a0-afbd-f88359db6324" (UID: "389d6832-2d3e-49a0-afbd-f88359db6324"). InnerVolumeSpecName "kube-api-access-bvk6s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:05:45 crc kubenswrapper[4787]: I1001 10:05:45.179666 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/389d6832-2d3e-49a0-afbd-f88359db6324-inventory" (OuterVolumeSpecName: "inventory") pod "389d6832-2d3e-49a0-afbd-f88359db6324" (UID: "389d6832-2d3e-49a0-afbd-f88359db6324"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:05:45 crc kubenswrapper[4787]: I1001 10:05:45.187812 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/389d6832-2d3e-49a0-afbd-f88359db6324-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "389d6832-2d3e-49a0-afbd-f88359db6324" (UID: "389d6832-2d3e-49a0-afbd-f88359db6324"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:05:45 crc kubenswrapper[4787]: I1001 10:05:45.251328 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bvk6s\" (UniqueName: \"kubernetes.io/projected/389d6832-2d3e-49a0-afbd-f88359db6324-kube-api-access-bvk6s\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:45 crc kubenswrapper[4787]: I1001 10:05:45.251655 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/389d6832-2d3e-49a0-afbd-f88359db6324-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:45 crc kubenswrapper[4787]: I1001 10:05:45.251665 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/389d6832-2d3e-49a0-afbd-f88359db6324-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:45 crc kubenswrapper[4787]: I1001 10:05:45.524613 4787 scope.go:117] "RemoveContainer" containerID="e01333312cf21546b537b606d139d2d34e67fa8885c08f9ceb1f4c790ae4fd3a" Oct 01 10:05:45 crc kubenswrapper[4787]: I1001 10:05:45.715195 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ws2cx" event={"ID":"389d6832-2d3e-49a0-afbd-f88359db6324","Type":"ContainerDied","Data":"4388f421744e442f6afd5c31d56eac8753241bbae9b25176a6f7582999009354"} Oct 01 10:05:45 crc kubenswrapper[4787]: I1001 10:05:45.715241 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ws2cx" Oct 01 10:05:45 crc kubenswrapper[4787]: I1001 10:05:45.715248 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4388f421744e442f6afd5c31d56eac8753241bbae9b25176a6f7582999009354" Oct 01 10:05:45 crc kubenswrapper[4787]: I1001 10:05:45.798785 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-llnrf"] Oct 01 10:05:45 crc kubenswrapper[4787]: E1001 10:05:45.799191 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="389d6832-2d3e-49a0-afbd-f88359db6324" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 01 10:05:45 crc kubenswrapper[4787]: I1001 10:05:45.799210 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="389d6832-2d3e-49a0-afbd-f88359db6324" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 01 10:05:45 crc kubenswrapper[4787]: I1001 10:05:45.799414 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="389d6832-2d3e-49a0-afbd-f88359db6324" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 01 10:05:45 crc kubenswrapper[4787]: I1001 10:05:45.800137 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-llnrf" Oct 01 10:05:45 crc kubenswrapper[4787]: I1001 10:05:45.803845 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 10:05:45 crc kubenswrapper[4787]: I1001 10:05:45.812364 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-llnrf"] Oct 01 10:05:45 crc kubenswrapper[4787]: I1001 10:05:45.812610 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 10:05:45 crc kubenswrapper[4787]: I1001 10:05:45.812796 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vh5v6" Oct 01 10:05:45 crc kubenswrapper[4787]: I1001 10:05:45.812923 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 10:05:45 crc kubenswrapper[4787]: I1001 10:05:45.864574 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/728521d2-1e71-4da3-a8bc-bd68e02eae35-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-llnrf\" (UID: \"728521d2-1e71-4da3-a8bc-bd68e02eae35\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-llnrf" Oct 01 10:05:45 crc kubenswrapper[4787]: I1001 10:05:45.865269 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/728521d2-1e71-4da3-a8bc-bd68e02eae35-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-llnrf\" (UID: \"728521d2-1e71-4da3-a8bc-bd68e02eae35\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-llnrf" Oct 01 10:05:45 crc kubenswrapper[4787]: I1001 10:05:45.873610 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llp4x\" (UniqueName: \"kubernetes.io/projected/728521d2-1e71-4da3-a8bc-bd68e02eae35-kube-api-access-llp4x\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-llnrf\" (UID: \"728521d2-1e71-4da3-a8bc-bd68e02eae35\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-llnrf" Oct 01 10:05:45 crc kubenswrapper[4787]: I1001 10:05:45.975540 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/728521d2-1e71-4da3-a8bc-bd68e02eae35-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-llnrf\" (UID: \"728521d2-1e71-4da3-a8bc-bd68e02eae35\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-llnrf" Oct 01 10:05:45 crc kubenswrapper[4787]: I1001 10:05:45.976276 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llp4x\" (UniqueName: \"kubernetes.io/projected/728521d2-1e71-4da3-a8bc-bd68e02eae35-kube-api-access-llp4x\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-llnrf\" (UID: \"728521d2-1e71-4da3-a8bc-bd68e02eae35\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-llnrf" Oct 01 10:05:45 crc kubenswrapper[4787]: I1001 10:05:45.976454 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/728521d2-1e71-4da3-a8bc-bd68e02eae35-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-llnrf\" (UID: \"728521d2-1e71-4da3-a8bc-bd68e02eae35\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-llnrf" Oct 01 10:05:45 crc kubenswrapper[4787]: I1001 10:05:45.983926 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/728521d2-1e71-4da3-a8bc-bd68e02eae35-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-llnrf\" (UID: \"728521d2-1e71-4da3-a8bc-bd68e02eae35\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-llnrf" Oct 01 10:05:45 crc kubenswrapper[4787]: I1001 10:05:45.985630 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/728521d2-1e71-4da3-a8bc-bd68e02eae35-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-llnrf\" (UID: \"728521d2-1e71-4da3-a8bc-bd68e02eae35\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-llnrf" Oct 01 10:05:45 crc kubenswrapper[4787]: I1001 10:05:45.994554 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llp4x\" (UniqueName: \"kubernetes.io/projected/728521d2-1e71-4da3-a8bc-bd68e02eae35-kube-api-access-llp4x\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-llnrf\" (UID: \"728521d2-1e71-4da3-a8bc-bd68e02eae35\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-llnrf" Oct 01 10:05:46 crc kubenswrapper[4787]: I1001 10:05:46.190463 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-llnrf" Oct 01 10:05:46 crc kubenswrapper[4787]: I1001 10:05:46.717487 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-llnrf"] Oct 01 10:05:46 crc kubenswrapper[4787]: W1001 10:05:46.729272 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod728521d2_1e71_4da3_a8bc_bd68e02eae35.slice/crio-af7370e201a9c99b9087583d62b1d17b6e2257d69d9233f3d70d42e5779d6fe1 WatchSource:0}: Error finding container af7370e201a9c99b9087583d62b1d17b6e2257d69d9233f3d70d42e5779d6fe1: Status 404 returned error can't find the container with id af7370e201a9c99b9087583d62b1d17b6e2257d69d9233f3d70d42e5779d6fe1 Oct 01 10:05:46 crc kubenswrapper[4787]: I1001 10:05:46.736050 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" event={"ID":"fa51b883-5f3e-4141-9d4c-37704aac7718","Type":"ContainerStarted","Data":"0044a9c9898ad2d4aaf6c39b3a1b6ec6368111794783cf724347932ef58f6d3d"} Oct 01 10:05:47 crc kubenswrapper[4787]: I1001 10:05:47.744986 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-llnrf" event={"ID":"728521d2-1e71-4da3-a8bc-bd68e02eae35","Type":"ContainerStarted","Data":"af7370e201a9c99b9087583d62b1d17b6e2257d69d9233f3d70d42e5779d6fe1"} Oct 01 10:05:48 crc kubenswrapper[4787]: I1001 10:05:48.756050 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-llnrf" event={"ID":"728521d2-1e71-4da3-a8bc-bd68e02eae35","Type":"ContainerStarted","Data":"8ec993aa38722a8a87df7289054c8316b01aeee9ec2d88d058bb23de7cf62fbb"} Oct 01 10:05:48 crc kubenswrapper[4787]: I1001 10:05:48.778346 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-llnrf" podStartSLOduration=2.505873103 podStartE2EDuration="3.778322994s" podCreationTimestamp="2025-10-01 10:05:45 +0000 UTC" firstStartedPulling="2025-10-01 10:05:46.731619818 +0000 UTC m=+1778.846763975" lastFinishedPulling="2025-10-01 10:05:48.004069709 +0000 UTC m=+1780.119213866" observedRunningTime="2025-10-01 10:05:48.770470939 +0000 UTC m=+1780.885615116" watchObservedRunningTime="2025-10-01 10:05:48.778322994 +0000 UTC m=+1780.893467161" Oct 01 10:05:57 crc kubenswrapper[4787]: I1001 10:05:57.841766 4787 generic.go:334] "Generic (PLEG): container finished" podID="728521d2-1e71-4da3-a8bc-bd68e02eae35" containerID="8ec993aa38722a8a87df7289054c8316b01aeee9ec2d88d058bb23de7cf62fbb" exitCode=0 Oct 01 10:05:57 crc kubenswrapper[4787]: I1001 10:05:57.841937 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-llnrf" event={"ID":"728521d2-1e71-4da3-a8bc-bd68e02eae35","Type":"ContainerDied","Data":"8ec993aa38722a8a87df7289054c8316b01aeee9ec2d88d058bb23de7cf62fbb"} Oct 01 10:05:59 crc kubenswrapper[4787]: I1001 10:05:59.238657 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-llnrf" Oct 01 10:05:59 crc kubenswrapper[4787]: I1001 10:05:59.357590 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llp4x\" (UniqueName: \"kubernetes.io/projected/728521d2-1e71-4da3-a8bc-bd68e02eae35-kube-api-access-llp4x\") pod \"728521d2-1e71-4da3-a8bc-bd68e02eae35\" (UID: \"728521d2-1e71-4da3-a8bc-bd68e02eae35\") " Oct 01 10:05:59 crc kubenswrapper[4787]: I1001 10:05:59.357962 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/728521d2-1e71-4da3-a8bc-bd68e02eae35-ssh-key\") pod \"728521d2-1e71-4da3-a8bc-bd68e02eae35\" (UID: \"728521d2-1e71-4da3-a8bc-bd68e02eae35\") " Oct 01 10:05:59 crc kubenswrapper[4787]: I1001 10:05:59.358132 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/728521d2-1e71-4da3-a8bc-bd68e02eae35-inventory\") pod \"728521d2-1e71-4da3-a8bc-bd68e02eae35\" (UID: \"728521d2-1e71-4da3-a8bc-bd68e02eae35\") " Oct 01 10:05:59 crc kubenswrapper[4787]: I1001 10:05:59.364862 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/728521d2-1e71-4da3-a8bc-bd68e02eae35-kube-api-access-llp4x" (OuterVolumeSpecName: "kube-api-access-llp4x") pod "728521d2-1e71-4da3-a8bc-bd68e02eae35" (UID: "728521d2-1e71-4da3-a8bc-bd68e02eae35"). InnerVolumeSpecName "kube-api-access-llp4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:05:59 crc kubenswrapper[4787]: I1001 10:05:59.383970 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/728521d2-1e71-4da3-a8bc-bd68e02eae35-inventory" (OuterVolumeSpecName: "inventory") pod "728521d2-1e71-4da3-a8bc-bd68e02eae35" (UID: "728521d2-1e71-4da3-a8bc-bd68e02eae35"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:05:59 crc kubenswrapper[4787]: I1001 10:05:59.386157 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/728521d2-1e71-4da3-a8bc-bd68e02eae35-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "728521d2-1e71-4da3-a8bc-bd68e02eae35" (UID: "728521d2-1e71-4da3-a8bc-bd68e02eae35"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:05:59 crc kubenswrapper[4787]: I1001 10:05:59.459708 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/728521d2-1e71-4da3-a8bc-bd68e02eae35-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:59 crc kubenswrapper[4787]: I1001 10:05:59.459735 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llp4x\" (UniqueName: \"kubernetes.io/projected/728521d2-1e71-4da3-a8bc-bd68e02eae35-kube-api-access-llp4x\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:59 crc kubenswrapper[4787]: I1001 10:05:59.459745 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/728521d2-1e71-4da3-a8bc-bd68e02eae35-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 10:05:59 crc kubenswrapper[4787]: I1001 10:05:59.869262 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-llnrf" event={"ID":"728521d2-1e71-4da3-a8bc-bd68e02eae35","Type":"ContainerDied","Data":"af7370e201a9c99b9087583d62b1d17b6e2257d69d9233f3d70d42e5779d6fe1"} Oct 01 10:05:59 crc kubenswrapper[4787]: I1001 10:05:59.869311 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af7370e201a9c99b9087583d62b1d17b6e2257d69d9233f3d70d42e5779d6fe1" Oct 01 10:05:59 crc kubenswrapper[4787]: I1001 10:05:59.869350 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-llnrf" Oct 01 10:05:59 crc kubenswrapper[4787]: I1001 10:05:59.940661 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv"] Oct 01 10:05:59 crc kubenswrapper[4787]: E1001 10:05:59.941234 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="728521d2-1e71-4da3-a8bc-bd68e02eae35" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 01 10:05:59 crc kubenswrapper[4787]: I1001 10:05:59.941326 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="728521d2-1e71-4da3-a8bc-bd68e02eae35" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 01 10:05:59 crc kubenswrapper[4787]: I1001 10:05:59.941603 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="728521d2-1e71-4da3-a8bc-bd68e02eae35" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 01 10:05:59 crc kubenswrapper[4787]: I1001 10:05:59.942417 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:05:59 crc kubenswrapper[4787]: I1001 10:05:59.944732 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Oct 01 10:05:59 crc kubenswrapper[4787]: I1001 10:05:59.945167 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Oct 01 10:05:59 crc kubenswrapper[4787]: I1001 10:05:59.945857 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 10:05:59 crc kubenswrapper[4787]: I1001 10:05:59.946032 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Oct 01 10:05:59 crc kubenswrapper[4787]: I1001 10:05:59.946291 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 10:05:59 crc kubenswrapper[4787]: I1001 10:05:59.946407 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vh5v6" Oct 01 10:05:59 crc kubenswrapper[4787]: I1001 10:05:59.946509 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 10:05:59 crc kubenswrapper[4787]: I1001 10:05:59.946615 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Oct 01 10:05:59 crc kubenswrapper[4787]: I1001 10:05:59.963670 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv"] Oct 01 10:05:59 crc kubenswrapper[4787]: I1001 10:05:59.970460 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:05:59 crc kubenswrapper[4787]: I1001 10:05:59.970531 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:05:59 crc kubenswrapper[4787]: I1001 10:05:59.970580 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:05:59 crc kubenswrapper[4787]: I1001 10:05:59.970616 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:05:59 crc kubenswrapper[4787]: I1001 10:05:59.970644 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6m8p\" (UniqueName: \"kubernetes.io/projected/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-kube-api-access-n6m8p\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:05:59 crc kubenswrapper[4787]: I1001 10:05:59.970686 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:05:59 crc kubenswrapper[4787]: I1001 10:05:59.970735 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:05:59 crc kubenswrapper[4787]: I1001 10:05:59.970758 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:05:59 crc kubenswrapper[4787]: I1001 10:05:59.970784 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:05:59 crc kubenswrapper[4787]: I1001 10:05:59.970819 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:05:59 crc kubenswrapper[4787]: I1001 10:05:59.970879 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:05:59 crc kubenswrapper[4787]: I1001 10:05:59.970905 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:05:59 crc kubenswrapper[4787]: I1001 10:05:59.970941 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:05:59 crc kubenswrapper[4787]: I1001 10:05:59.970966 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:06:00 crc kubenswrapper[4787]: I1001 10:06:00.072267 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:06:00 crc kubenswrapper[4787]: I1001 10:06:00.072328 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:06:00 crc kubenswrapper[4787]: I1001 10:06:00.072387 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:06:00 crc kubenswrapper[4787]: I1001 10:06:00.072423 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:06:00 crc kubenswrapper[4787]: I1001 10:06:00.072442 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:06:00 crc kubenswrapper[4787]: I1001 10:06:00.072472 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:06:00 crc kubenswrapper[4787]: I1001 10:06:00.072514 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:06:00 crc kubenswrapper[4787]: I1001 10:06:00.072544 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:06:00 crc kubenswrapper[4787]: I1001 10:06:00.072567 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:06:00 crc kubenswrapper[4787]: I1001 10:06:00.072591 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:06:00 crc kubenswrapper[4787]: I1001 10:06:00.072610 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6m8p\" (UniqueName: \"kubernetes.io/projected/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-kube-api-access-n6m8p\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:06:00 crc kubenswrapper[4787]: I1001 10:06:00.072642 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:06:00 crc kubenswrapper[4787]: I1001 10:06:00.072763 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:06:00 crc kubenswrapper[4787]: I1001 10:06:00.072783 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:06:00 crc kubenswrapper[4787]: I1001 10:06:00.076403 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:06:00 crc kubenswrapper[4787]: I1001 10:06:00.076704 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:06:00 crc kubenswrapper[4787]: I1001 10:06:00.077248 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:06:00 crc kubenswrapper[4787]: I1001 10:06:00.078373 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:06:00 crc kubenswrapper[4787]: I1001 10:06:00.078586 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:06:00 crc kubenswrapper[4787]: I1001 10:06:00.078756 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:06:00 crc kubenswrapper[4787]: I1001 10:06:00.078939 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:06:00 crc kubenswrapper[4787]: I1001 10:06:00.079564 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:06:00 crc kubenswrapper[4787]: I1001 10:06:00.080002 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:06:00 crc kubenswrapper[4787]: I1001 10:06:00.080377 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:06:00 crc kubenswrapper[4787]: I1001 10:06:00.080671 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:06:00 crc kubenswrapper[4787]: I1001 10:06:00.081065 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:06:00 crc kubenswrapper[4787]: I1001 10:06:00.086119 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:06:00 crc kubenswrapper[4787]: I1001 10:06:00.091158 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6m8p\" (UniqueName: \"kubernetes.io/projected/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-kube-api-access-n6m8p\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-79skv\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:06:00 crc kubenswrapper[4787]: I1001 10:06:00.271205 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:06:00 crc kubenswrapper[4787]: I1001 10:06:00.778302 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv"] Oct 01 10:06:00 crc kubenswrapper[4787]: I1001 10:06:00.888306 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" event={"ID":"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1","Type":"ContainerStarted","Data":"6640f7f7845b5dda4e566c0235559fbef458bf9050fd6229b4fc7cb2007c0bf7"} Oct 01 10:06:01 crc kubenswrapper[4787]: I1001 10:06:01.897697 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" event={"ID":"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1","Type":"ContainerStarted","Data":"e933eeac500497311ac8226e728da5751c90d12ae9d6cdfcb06d51cf06abb427"} Oct 01 10:06:01 crc kubenswrapper[4787]: I1001 10:06:01.939119 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" podStartSLOduration=2.365218521 podStartE2EDuration="2.939063829s" podCreationTimestamp="2025-10-01 10:05:59 +0000 UTC" firstStartedPulling="2025-10-01 10:06:00.786274295 +0000 UTC m=+1792.901418462" lastFinishedPulling="2025-10-01 10:06:01.360119613 +0000 UTC m=+1793.475263770" observedRunningTime="2025-10-01 10:06:01.926581168 +0000 UTC m=+1794.041725355" watchObservedRunningTime="2025-10-01 10:06:01.939063829 +0000 UTC m=+1794.054208016" Oct 01 10:06:40 crc kubenswrapper[4787]: I1001 10:06:40.287755 4787 generic.go:334] "Generic (PLEG): container finished" podID="d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1" containerID="e933eeac500497311ac8226e728da5751c90d12ae9d6cdfcb06d51cf06abb427" exitCode=0 Oct 01 10:06:40 crc kubenswrapper[4787]: I1001 10:06:40.287835 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" event={"ID":"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1","Type":"ContainerDied","Data":"e933eeac500497311ac8226e728da5751c90d12ae9d6cdfcb06d51cf06abb427"} Oct 01 10:06:41 crc kubenswrapper[4787]: I1001 10:06:41.830522 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:06:41 crc kubenswrapper[4787]: I1001 10:06:41.906007 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-openstack-edpm-ipam-ovn-default-certs-0\") pod \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " Oct 01 10:06:41 crc kubenswrapper[4787]: I1001 10:06:41.906097 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-neutron-metadata-combined-ca-bundle\") pod \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " Oct 01 10:06:41 crc kubenswrapper[4787]: I1001 10:06:41.906226 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-ssh-key\") pod \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " Oct 01 10:06:41 crc kubenswrapper[4787]: I1001 10:06:41.906296 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " Oct 01 10:06:41 crc kubenswrapper[4787]: I1001 10:06:41.906373 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6m8p\" (UniqueName: \"kubernetes.io/projected/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-kube-api-access-n6m8p\") pod \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " Oct 01 10:06:41 crc kubenswrapper[4787]: I1001 10:06:41.906444 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-ovn-combined-ca-bundle\") pod \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " Oct 01 10:06:41 crc kubenswrapper[4787]: I1001 10:06:41.906490 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " Oct 01 10:06:41 crc kubenswrapper[4787]: I1001 10:06:41.906542 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-telemetry-combined-ca-bundle\") pod \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " Oct 01 10:06:41 crc kubenswrapper[4787]: I1001 10:06:41.906572 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-nova-combined-ca-bundle\") pod \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " Oct 01 10:06:41 crc kubenswrapper[4787]: I1001 10:06:41.906602 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " Oct 01 10:06:41 crc kubenswrapper[4787]: I1001 10:06:41.906657 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-bootstrap-combined-ca-bundle\") pod \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " Oct 01 10:06:41 crc kubenswrapper[4787]: I1001 10:06:41.906698 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-libvirt-combined-ca-bundle\") pod \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " Oct 01 10:06:41 crc kubenswrapper[4787]: I1001 10:06:41.906756 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-inventory\") pod \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " Oct 01 10:06:41 crc kubenswrapper[4787]: I1001 10:06:41.906827 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-repo-setup-combined-ca-bundle\") pod \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\" (UID: \"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1\") " Oct 01 10:06:41 crc kubenswrapper[4787]: I1001 10:06:41.917526 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1" (UID: "d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:06:41 crc kubenswrapper[4787]: I1001 10:06:41.917679 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1" (UID: "d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:06:41 crc kubenswrapper[4787]: I1001 10:06:41.918558 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1" (UID: "d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:06:41 crc kubenswrapper[4787]: I1001 10:06:41.925446 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1" (UID: "d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:06:41 crc kubenswrapper[4787]: I1001 10:06:41.925532 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1" (UID: "d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:06:41 crc kubenswrapper[4787]: I1001 10:06:41.925614 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1" (UID: "d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:06:41 crc kubenswrapper[4787]: I1001 10:06:41.925686 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1" (UID: "d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:06:41 crc kubenswrapper[4787]: I1001 10:06:41.926030 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1" (UID: "d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:06:41 crc kubenswrapper[4787]: I1001 10:06:41.926447 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-kube-api-access-n6m8p" (OuterVolumeSpecName: "kube-api-access-n6m8p") pod "d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1" (UID: "d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1"). InnerVolumeSpecName "kube-api-access-n6m8p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:06:41 crc kubenswrapper[4787]: I1001 10:06:41.929192 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1" (UID: "d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:06:41 crc kubenswrapper[4787]: I1001 10:06:41.932502 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1" (UID: "d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:06:41 crc kubenswrapper[4787]: I1001 10:06:41.932865 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1" (UID: "d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:06:41 crc kubenswrapper[4787]: I1001 10:06:41.953849 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1" (UID: "d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:06:41 crc kubenswrapper[4787]: I1001 10:06:41.960790 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-inventory" (OuterVolumeSpecName: "inventory") pod "d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1" (UID: "d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.009653 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6m8p\" (UniqueName: \"kubernetes.io/projected/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-kube-api-access-n6m8p\") on node \"crc\" DevicePath \"\"" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.009691 4787 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.009704 4787 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.009715 4787 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.009725 4787 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.009737 4787 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.009748 4787 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.009757 4787 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.009769 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.009777 4787 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.009785 4787 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.009797 4787 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.009805 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.009814 4787 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.321110 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" event={"ID":"d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1","Type":"ContainerDied","Data":"6640f7f7845b5dda4e566c0235559fbef458bf9050fd6229b4fc7cb2007c0bf7"} Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.321423 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6640f7f7845b5dda4e566c0235559fbef458bf9050fd6229b4fc7cb2007c0bf7" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.321268 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-79skv" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.444135 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-htlbc"] Oct 01 10:06:42 crc kubenswrapper[4787]: E1001 10:06:42.444543 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.444564 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.444820 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.445911 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-htlbc" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.448254 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.449255 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vh5v6" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.449721 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.449726 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.452759 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.481475 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-htlbc"] Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.522592 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dc18c70f-f7c7-4a60-87e6-699320d382fc-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-htlbc\" (UID: \"dc18c70f-f7c7-4a60-87e6-699320d382fc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-htlbc" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.522726 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc18c70f-f7c7-4a60-87e6-699320d382fc-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-htlbc\" (UID: \"dc18c70f-f7c7-4a60-87e6-699320d382fc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-htlbc" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.522964 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/dc18c70f-f7c7-4a60-87e6-699320d382fc-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-htlbc\" (UID: \"dc18c70f-f7c7-4a60-87e6-699320d382fc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-htlbc" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.523404 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dc18c70f-f7c7-4a60-87e6-699320d382fc-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-htlbc\" (UID: \"dc18c70f-f7c7-4a60-87e6-699320d382fc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-htlbc" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.523584 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpcv7\" (UniqueName: \"kubernetes.io/projected/dc18c70f-f7c7-4a60-87e6-699320d382fc-kube-api-access-jpcv7\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-htlbc\" (UID: \"dc18c70f-f7c7-4a60-87e6-699320d382fc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-htlbc" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.627146 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dc18c70f-f7c7-4a60-87e6-699320d382fc-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-htlbc\" (UID: \"dc18c70f-f7c7-4a60-87e6-699320d382fc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-htlbc" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.627253 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpcv7\" (UniqueName: \"kubernetes.io/projected/dc18c70f-f7c7-4a60-87e6-699320d382fc-kube-api-access-jpcv7\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-htlbc\" (UID: \"dc18c70f-f7c7-4a60-87e6-699320d382fc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-htlbc" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.627316 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dc18c70f-f7c7-4a60-87e6-699320d382fc-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-htlbc\" (UID: \"dc18c70f-f7c7-4a60-87e6-699320d382fc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-htlbc" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.627443 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc18c70f-f7c7-4a60-87e6-699320d382fc-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-htlbc\" (UID: \"dc18c70f-f7c7-4a60-87e6-699320d382fc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-htlbc" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.627523 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/dc18c70f-f7c7-4a60-87e6-699320d382fc-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-htlbc\" (UID: \"dc18c70f-f7c7-4a60-87e6-699320d382fc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-htlbc" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.631504 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/dc18c70f-f7c7-4a60-87e6-699320d382fc-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-htlbc\" (UID: \"dc18c70f-f7c7-4a60-87e6-699320d382fc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-htlbc" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.636194 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dc18c70f-f7c7-4a60-87e6-699320d382fc-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-htlbc\" (UID: \"dc18c70f-f7c7-4a60-87e6-699320d382fc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-htlbc" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.636430 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dc18c70f-f7c7-4a60-87e6-699320d382fc-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-htlbc\" (UID: \"dc18c70f-f7c7-4a60-87e6-699320d382fc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-htlbc" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.649496 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc18c70f-f7c7-4a60-87e6-699320d382fc-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-htlbc\" (UID: \"dc18c70f-f7c7-4a60-87e6-699320d382fc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-htlbc" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.651227 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpcv7\" (UniqueName: \"kubernetes.io/projected/dc18c70f-f7c7-4a60-87e6-699320d382fc-kube-api-access-jpcv7\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-htlbc\" (UID: \"dc18c70f-f7c7-4a60-87e6-699320d382fc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-htlbc" Oct 01 10:06:42 crc kubenswrapper[4787]: I1001 10:06:42.765989 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-htlbc" Oct 01 10:06:43 crc kubenswrapper[4787]: I1001 10:06:43.395127 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-htlbc"] Oct 01 10:06:43 crc kubenswrapper[4787]: I1001 10:06:43.398504 4787 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 10:06:44 crc kubenswrapper[4787]: I1001 10:06:44.355987 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-htlbc" event={"ID":"dc18c70f-f7c7-4a60-87e6-699320d382fc","Type":"ContainerStarted","Data":"80537c40129a2f364c4dc56f0294e69584b4c1e62efd9e7fbdd5077ebd5b7109"} Oct 01 10:06:44 crc kubenswrapper[4787]: I1001 10:06:44.356568 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-htlbc" event={"ID":"dc18c70f-f7c7-4a60-87e6-699320d382fc","Type":"ContainerStarted","Data":"f408ac17b1f4ebf5f1fccafde3ac5d64cc90cc12130bde2b70c34ec04866b93e"} Oct 01 10:06:44 crc kubenswrapper[4787]: I1001 10:06:44.384050 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-htlbc" podStartSLOduration=1.8210112920000001 podStartE2EDuration="2.384027561s" podCreationTimestamp="2025-10-01 10:06:42 +0000 UTC" firstStartedPulling="2025-10-01 10:06:43.398011756 +0000 UTC m=+1835.513155943" lastFinishedPulling="2025-10-01 10:06:43.961028055 +0000 UTC m=+1836.076172212" observedRunningTime="2025-10-01 10:06:44.37716076 +0000 UTC m=+1836.492304917" watchObservedRunningTime="2025-10-01 10:06:44.384027561 +0000 UTC m=+1836.499171718" Oct 01 10:06:58 crc kubenswrapper[4787]: I1001 10:06:58.380244 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-5c7b59dd4f-spxm4" podUID="e743cd5e-e8c6-4fe2-9480-3a30316b8e23" containerName="proxy-server" probeResult="failure" output="HTTP probe failed with statuscode: 502" Oct 01 10:07:47 crc kubenswrapper[4787]: I1001 10:07:47.931665 4787 generic.go:334] "Generic (PLEG): container finished" podID="dc18c70f-f7c7-4a60-87e6-699320d382fc" containerID="80537c40129a2f364c4dc56f0294e69584b4c1e62efd9e7fbdd5077ebd5b7109" exitCode=0 Oct 01 10:07:47 crc kubenswrapper[4787]: I1001 10:07:47.931737 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-htlbc" event={"ID":"dc18c70f-f7c7-4a60-87e6-699320d382fc","Type":"ContainerDied","Data":"80537c40129a2f364c4dc56f0294e69584b4c1e62efd9e7fbdd5077ebd5b7109"} Oct 01 10:07:49 crc kubenswrapper[4787]: I1001 10:07:49.340678 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-htlbc" Oct 01 10:07:49 crc kubenswrapper[4787]: I1001 10:07:49.461548 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dc18c70f-f7c7-4a60-87e6-699320d382fc-ssh-key\") pod \"dc18c70f-f7c7-4a60-87e6-699320d382fc\" (UID: \"dc18c70f-f7c7-4a60-87e6-699320d382fc\") " Oct 01 10:07:49 crc kubenswrapper[4787]: I1001 10:07:49.461683 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jpcv7\" (UniqueName: \"kubernetes.io/projected/dc18c70f-f7c7-4a60-87e6-699320d382fc-kube-api-access-jpcv7\") pod \"dc18c70f-f7c7-4a60-87e6-699320d382fc\" (UID: \"dc18c70f-f7c7-4a60-87e6-699320d382fc\") " Oct 01 10:07:49 crc kubenswrapper[4787]: I1001 10:07:49.461726 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc18c70f-f7c7-4a60-87e6-699320d382fc-ovn-combined-ca-bundle\") pod \"dc18c70f-f7c7-4a60-87e6-699320d382fc\" (UID: \"dc18c70f-f7c7-4a60-87e6-699320d382fc\") " Oct 01 10:07:49 crc kubenswrapper[4787]: I1001 10:07:49.461779 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/dc18c70f-f7c7-4a60-87e6-699320d382fc-ovncontroller-config-0\") pod \"dc18c70f-f7c7-4a60-87e6-699320d382fc\" (UID: \"dc18c70f-f7c7-4a60-87e6-699320d382fc\") " Oct 01 10:07:49 crc kubenswrapper[4787]: I1001 10:07:49.461864 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dc18c70f-f7c7-4a60-87e6-699320d382fc-inventory\") pod \"dc18c70f-f7c7-4a60-87e6-699320d382fc\" (UID: \"dc18c70f-f7c7-4a60-87e6-699320d382fc\") " Oct 01 10:07:49 crc kubenswrapper[4787]: I1001 10:07:49.467378 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc18c70f-f7c7-4a60-87e6-699320d382fc-kube-api-access-jpcv7" (OuterVolumeSpecName: "kube-api-access-jpcv7") pod "dc18c70f-f7c7-4a60-87e6-699320d382fc" (UID: "dc18c70f-f7c7-4a60-87e6-699320d382fc"). InnerVolumeSpecName "kube-api-access-jpcv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:07:49 crc kubenswrapper[4787]: I1001 10:07:49.467613 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc18c70f-f7c7-4a60-87e6-699320d382fc-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "dc18c70f-f7c7-4a60-87e6-699320d382fc" (UID: "dc18c70f-f7c7-4a60-87e6-699320d382fc"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:07:49 crc kubenswrapper[4787]: I1001 10:07:49.493474 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc18c70f-f7c7-4a60-87e6-699320d382fc-inventory" (OuterVolumeSpecName: "inventory") pod "dc18c70f-f7c7-4a60-87e6-699320d382fc" (UID: "dc18c70f-f7c7-4a60-87e6-699320d382fc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:07:49 crc kubenswrapper[4787]: I1001 10:07:49.495723 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc18c70f-f7c7-4a60-87e6-699320d382fc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "dc18c70f-f7c7-4a60-87e6-699320d382fc" (UID: "dc18c70f-f7c7-4a60-87e6-699320d382fc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:07:49 crc kubenswrapper[4787]: I1001 10:07:49.499311 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc18c70f-f7c7-4a60-87e6-699320d382fc-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "dc18c70f-f7c7-4a60-87e6-699320d382fc" (UID: "dc18c70f-f7c7-4a60-87e6-699320d382fc"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:07:49 crc kubenswrapper[4787]: I1001 10:07:49.578641 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jpcv7\" (UniqueName: \"kubernetes.io/projected/dc18c70f-f7c7-4a60-87e6-699320d382fc-kube-api-access-jpcv7\") on node \"crc\" DevicePath \"\"" Oct 01 10:07:49 crc kubenswrapper[4787]: I1001 10:07:49.578689 4787 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc18c70f-f7c7-4a60-87e6-699320d382fc-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 10:07:49 crc kubenswrapper[4787]: I1001 10:07:49.578704 4787 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/dc18c70f-f7c7-4a60-87e6-699320d382fc-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 01 10:07:49 crc kubenswrapper[4787]: I1001 10:07:49.578721 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dc18c70f-f7c7-4a60-87e6-699320d382fc-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 10:07:49 crc kubenswrapper[4787]: I1001 10:07:49.578732 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dc18c70f-f7c7-4a60-87e6-699320d382fc-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 10:07:49 crc kubenswrapper[4787]: I1001 10:07:49.949443 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-htlbc" event={"ID":"dc18c70f-f7c7-4a60-87e6-699320d382fc","Type":"ContainerDied","Data":"f408ac17b1f4ebf5f1fccafde3ac5d64cc90cc12130bde2b70c34ec04866b93e"} Oct 01 10:07:49 crc kubenswrapper[4787]: I1001 10:07:49.949494 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f408ac17b1f4ebf5f1fccafde3ac5d64cc90cc12130bde2b70c34ec04866b93e" Oct 01 10:07:49 crc kubenswrapper[4787]: I1001 10:07:49.949527 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-htlbc" Oct 01 10:07:50 crc kubenswrapper[4787]: I1001 10:07:50.100594 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6"] Oct 01 10:07:50 crc kubenswrapper[4787]: E1001 10:07:50.100948 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc18c70f-f7c7-4a60-87e6-699320d382fc" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 01 10:07:50 crc kubenswrapper[4787]: I1001 10:07:50.100964 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc18c70f-f7c7-4a60-87e6-699320d382fc" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 01 10:07:50 crc kubenswrapper[4787]: I1001 10:07:50.101194 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc18c70f-f7c7-4a60-87e6-699320d382fc" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 01 10:07:50 crc kubenswrapper[4787]: I1001 10:07:50.101856 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6" Oct 01 10:07:50 crc kubenswrapper[4787]: I1001 10:07:50.104766 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 01 10:07:50 crc kubenswrapper[4787]: I1001 10:07:50.104854 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 01 10:07:50 crc kubenswrapper[4787]: I1001 10:07:50.105332 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 10:07:50 crc kubenswrapper[4787]: I1001 10:07:50.105530 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 10:07:50 crc kubenswrapper[4787]: I1001 10:07:50.105576 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vh5v6" Oct 01 10:07:50 crc kubenswrapper[4787]: I1001 10:07:50.107834 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 10:07:50 crc kubenswrapper[4787]: I1001 10:07:50.119502 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6"] Oct 01 10:07:50 crc kubenswrapper[4787]: I1001 10:07:50.291990 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e7fb97de-2759-41c1-b090-99ed2c95e92c-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6\" (UID: \"e7fb97de-2759-41c1-b090-99ed2c95e92c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6" Oct 01 10:07:50 crc kubenswrapper[4787]: I1001 10:07:50.292166 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drtgh\" (UniqueName: \"kubernetes.io/projected/e7fb97de-2759-41c1-b090-99ed2c95e92c-kube-api-access-drtgh\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6\" (UID: \"e7fb97de-2759-41c1-b090-99ed2c95e92c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6" Oct 01 10:07:50 crc kubenswrapper[4787]: I1001 10:07:50.292267 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e7fb97de-2759-41c1-b090-99ed2c95e92c-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6\" (UID: \"e7fb97de-2759-41c1-b090-99ed2c95e92c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6" Oct 01 10:07:50 crc kubenswrapper[4787]: I1001 10:07:50.292487 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e7fb97de-2759-41c1-b090-99ed2c95e92c-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6\" (UID: \"e7fb97de-2759-41c1-b090-99ed2c95e92c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6" Oct 01 10:07:50 crc kubenswrapper[4787]: I1001 10:07:50.292674 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e7fb97de-2759-41c1-b090-99ed2c95e92c-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6\" (UID: \"e7fb97de-2759-41c1-b090-99ed2c95e92c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6" Oct 01 10:07:50 crc kubenswrapper[4787]: I1001 10:07:50.292719 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7fb97de-2759-41c1-b090-99ed2c95e92c-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6\" (UID: \"e7fb97de-2759-41c1-b090-99ed2c95e92c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6" Oct 01 10:07:50 crc kubenswrapper[4787]: I1001 10:07:50.396587 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e7fb97de-2759-41c1-b090-99ed2c95e92c-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6\" (UID: \"e7fb97de-2759-41c1-b090-99ed2c95e92c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6" Oct 01 10:07:50 crc kubenswrapper[4787]: I1001 10:07:50.396765 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e7fb97de-2759-41c1-b090-99ed2c95e92c-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6\" (UID: \"e7fb97de-2759-41c1-b090-99ed2c95e92c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6" Oct 01 10:07:50 crc kubenswrapper[4787]: I1001 10:07:50.396814 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7fb97de-2759-41c1-b090-99ed2c95e92c-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6\" (UID: \"e7fb97de-2759-41c1-b090-99ed2c95e92c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6" Oct 01 10:07:50 crc kubenswrapper[4787]: I1001 10:07:50.397011 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e7fb97de-2759-41c1-b090-99ed2c95e92c-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6\" (UID: \"e7fb97de-2759-41c1-b090-99ed2c95e92c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6" Oct 01 10:07:50 crc kubenswrapper[4787]: I1001 10:07:50.397219 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drtgh\" (UniqueName: \"kubernetes.io/projected/e7fb97de-2759-41c1-b090-99ed2c95e92c-kube-api-access-drtgh\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6\" (UID: \"e7fb97de-2759-41c1-b090-99ed2c95e92c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6" Oct 01 10:07:50 crc kubenswrapper[4787]: I1001 10:07:50.397356 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e7fb97de-2759-41c1-b090-99ed2c95e92c-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6\" (UID: \"e7fb97de-2759-41c1-b090-99ed2c95e92c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6" Oct 01 10:07:50 crc kubenswrapper[4787]: I1001 10:07:50.402046 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e7fb97de-2759-41c1-b090-99ed2c95e92c-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6\" (UID: \"e7fb97de-2759-41c1-b090-99ed2c95e92c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6" Oct 01 10:07:50 crc kubenswrapper[4787]: I1001 10:07:50.402297 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e7fb97de-2759-41c1-b090-99ed2c95e92c-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6\" (UID: \"e7fb97de-2759-41c1-b090-99ed2c95e92c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6" Oct 01 10:07:50 crc kubenswrapper[4787]: I1001 10:07:50.402945 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7fb97de-2759-41c1-b090-99ed2c95e92c-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6\" (UID: \"e7fb97de-2759-41c1-b090-99ed2c95e92c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6" Oct 01 10:07:50 crc kubenswrapper[4787]: I1001 10:07:50.410965 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e7fb97de-2759-41c1-b090-99ed2c95e92c-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6\" (UID: \"e7fb97de-2759-41c1-b090-99ed2c95e92c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6" Oct 01 10:07:50 crc kubenswrapper[4787]: I1001 10:07:50.411555 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e7fb97de-2759-41c1-b090-99ed2c95e92c-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6\" (UID: \"e7fb97de-2759-41c1-b090-99ed2c95e92c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6" Oct 01 10:07:50 crc kubenswrapper[4787]: I1001 10:07:50.414800 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drtgh\" (UniqueName: \"kubernetes.io/projected/e7fb97de-2759-41c1-b090-99ed2c95e92c-kube-api-access-drtgh\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6\" (UID: \"e7fb97de-2759-41c1-b090-99ed2c95e92c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6" Oct 01 10:07:50 crc kubenswrapper[4787]: I1001 10:07:50.433481 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6" Oct 01 10:07:50 crc kubenswrapper[4787]: I1001 10:07:50.977757 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6"] Oct 01 10:07:51 crc kubenswrapper[4787]: I1001 10:07:51.972689 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6" event={"ID":"e7fb97de-2759-41c1-b090-99ed2c95e92c","Type":"ContainerStarted","Data":"d95a244912a38ecfa0912efbc8fec6ce9ee1852e0e782a7fa6a3921a0e023ab0"} Oct 01 10:07:52 crc kubenswrapper[4787]: I1001 10:07:52.983291 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6" event={"ID":"e7fb97de-2759-41c1-b090-99ed2c95e92c","Type":"ContainerStarted","Data":"5a35f4c8d85427983b1ead48ba5031394e943d3a32f28afc3e0a44f4ec8ed474"} Oct 01 10:07:53 crc kubenswrapper[4787]: I1001 10:07:53.000353 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6" podStartSLOduration=1.9181509 podStartE2EDuration="3.000333445s" podCreationTimestamp="2025-10-01 10:07:50 +0000 UTC" firstStartedPulling="2025-10-01 10:07:50.986728018 +0000 UTC m=+1903.101872175" lastFinishedPulling="2025-10-01 10:07:52.068910523 +0000 UTC m=+1904.184054720" observedRunningTime="2025-10-01 10:07:52.998340686 +0000 UTC m=+1905.113484853" watchObservedRunningTime="2025-10-01 10:07:53.000333445 +0000 UTC m=+1905.115477602" Oct 01 10:08:11 crc kubenswrapper[4787]: I1001 10:08:11.250905 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:08:11 crc kubenswrapper[4787]: I1001 10:08:11.252350 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:08:38 crc kubenswrapper[4787]: I1001 10:08:38.405807 4787 generic.go:334] "Generic (PLEG): container finished" podID="e7fb97de-2759-41c1-b090-99ed2c95e92c" containerID="5a35f4c8d85427983b1ead48ba5031394e943d3a32f28afc3e0a44f4ec8ed474" exitCode=0 Oct 01 10:08:38 crc kubenswrapper[4787]: I1001 10:08:38.406222 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6" event={"ID":"e7fb97de-2759-41c1-b090-99ed2c95e92c","Type":"ContainerDied","Data":"5a35f4c8d85427983b1ead48ba5031394e943d3a32f28afc3e0a44f4ec8ed474"} Oct 01 10:08:39 crc kubenswrapper[4787]: I1001 10:08:39.850991 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6" Oct 01 10:08:39 crc kubenswrapper[4787]: I1001 10:08:39.973011 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7fb97de-2759-41c1-b090-99ed2c95e92c-neutron-metadata-combined-ca-bundle\") pod \"e7fb97de-2759-41c1-b090-99ed2c95e92c\" (UID: \"e7fb97de-2759-41c1-b090-99ed2c95e92c\") " Oct 01 10:08:39 crc kubenswrapper[4787]: I1001 10:08:39.973220 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-drtgh\" (UniqueName: \"kubernetes.io/projected/e7fb97de-2759-41c1-b090-99ed2c95e92c-kube-api-access-drtgh\") pod \"e7fb97de-2759-41c1-b090-99ed2c95e92c\" (UID: \"e7fb97de-2759-41c1-b090-99ed2c95e92c\") " Oct 01 10:08:39 crc kubenswrapper[4787]: I1001 10:08:39.973278 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e7fb97de-2759-41c1-b090-99ed2c95e92c-inventory\") pod \"e7fb97de-2759-41c1-b090-99ed2c95e92c\" (UID: \"e7fb97de-2759-41c1-b090-99ed2c95e92c\") " Oct 01 10:08:39 crc kubenswrapper[4787]: I1001 10:08:39.973315 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e7fb97de-2759-41c1-b090-99ed2c95e92c-neutron-ovn-metadata-agent-neutron-config-0\") pod \"e7fb97de-2759-41c1-b090-99ed2c95e92c\" (UID: \"e7fb97de-2759-41c1-b090-99ed2c95e92c\") " Oct 01 10:08:39 crc kubenswrapper[4787]: I1001 10:08:39.973368 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e7fb97de-2759-41c1-b090-99ed2c95e92c-ssh-key\") pod \"e7fb97de-2759-41c1-b090-99ed2c95e92c\" (UID: \"e7fb97de-2759-41c1-b090-99ed2c95e92c\") " Oct 01 10:08:39 crc kubenswrapper[4787]: I1001 10:08:39.973387 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e7fb97de-2759-41c1-b090-99ed2c95e92c-nova-metadata-neutron-config-0\") pod \"e7fb97de-2759-41c1-b090-99ed2c95e92c\" (UID: \"e7fb97de-2759-41c1-b090-99ed2c95e92c\") " Oct 01 10:08:39 crc kubenswrapper[4787]: I1001 10:08:39.985235 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7fb97de-2759-41c1-b090-99ed2c95e92c-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "e7fb97de-2759-41c1-b090-99ed2c95e92c" (UID: "e7fb97de-2759-41c1-b090-99ed2c95e92c"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:08:39 crc kubenswrapper[4787]: I1001 10:08:39.985251 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7fb97de-2759-41c1-b090-99ed2c95e92c-kube-api-access-drtgh" (OuterVolumeSpecName: "kube-api-access-drtgh") pod "e7fb97de-2759-41c1-b090-99ed2c95e92c" (UID: "e7fb97de-2759-41c1-b090-99ed2c95e92c"). InnerVolumeSpecName "kube-api-access-drtgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:08:40 crc kubenswrapper[4787]: I1001 10:08:40.003409 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7fb97de-2759-41c1-b090-99ed2c95e92c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e7fb97de-2759-41c1-b090-99ed2c95e92c" (UID: "e7fb97de-2759-41c1-b090-99ed2c95e92c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:08:40 crc kubenswrapper[4787]: I1001 10:08:40.004237 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7fb97de-2759-41c1-b090-99ed2c95e92c-inventory" (OuterVolumeSpecName: "inventory") pod "e7fb97de-2759-41c1-b090-99ed2c95e92c" (UID: "e7fb97de-2759-41c1-b090-99ed2c95e92c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:08:40 crc kubenswrapper[4787]: I1001 10:08:40.005312 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7fb97de-2759-41c1-b090-99ed2c95e92c-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "e7fb97de-2759-41c1-b090-99ed2c95e92c" (UID: "e7fb97de-2759-41c1-b090-99ed2c95e92c"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:08:40 crc kubenswrapper[4787]: I1001 10:08:40.007317 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7fb97de-2759-41c1-b090-99ed2c95e92c-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "e7fb97de-2759-41c1-b090-99ed2c95e92c" (UID: "e7fb97de-2759-41c1-b090-99ed2c95e92c"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:08:40 crc kubenswrapper[4787]: I1001 10:08:40.075641 4787 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7fb97de-2759-41c1-b090-99ed2c95e92c-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 10:08:40 crc kubenswrapper[4787]: I1001 10:08:40.075678 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-drtgh\" (UniqueName: \"kubernetes.io/projected/e7fb97de-2759-41c1-b090-99ed2c95e92c-kube-api-access-drtgh\") on node \"crc\" DevicePath \"\"" Oct 01 10:08:40 crc kubenswrapper[4787]: I1001 10:08:40.075690 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e7fb97de-2759-41c1-b090-99ed2c95e92c-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 10:08:40 crc kubenswrapper[4787]: I1001 10:08:40.075700 4787 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e7fb97de-2759-41c1-b090-99ed2c95e92c-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 01 10:08:40 crc kubenswrapper[4787]: I1001 10:08:40.075711 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e7fb97de-2759-41c1-b090-99ed2c95e92c-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 10:08:40 crc kubenswrapper[4787]: I1001 10:08:40.075720 4787 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e7fb97de-2759-41c1-b090-99ed2c95e92c-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 01 10:08:40 crc kubenswrapper[4787]: I1001 10:08:40.425050 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6" event={"ID":"e7fb97de-2759-41c1-b090-99ed2c95e92c","Type":"ContainerDied","Data":"d95a244912a38ecfa0912efbc8fec6ce9ee1852e0e782a7fa6a3921a0e023ab0"} Oct 01 10:08:40 crc kubenswrapper[4787]: I1001 10:08:40.425117 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6" Oct 01 10:08:40 crc kubenswrapper[4787]: I1001 10:08:40.425131 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d95a244912a38ecfa0912efbc8fec6ce9ee1852e0e782a7fa6a3921a0e023ab0" Oct 01 10:08:40 crc kubenswrapper[4787]: I1001 10:08:40.513266 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr"] Oct 01 10:08:40 crc kubenswrapper[4787]: E1001 10:08:40.513788 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7fb97de-2759-41c1-b090-99ed2c95e92c" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 01 10:08:40 crc kubenswrapper[4787]: I1001 10:08:40.513824 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7fb97de-2759-41c1-b090-99ed2c95e92c" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 01 10:08:40 crc kubenswrapper[4787]: I1001 10:08:40.514171 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7fb97de-2759-41c1-b090-99ed2c95e92c" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 01 10:08:40 crc kubenswrapper[4787]: I1001 10:08:40.515168 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr" Oct 01 10:08:40 crc kubenswrapper[4787]: I1001 10:08:40.518611 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 10:08:40 crc kubenswrapper[4787]: I1001 10:08:40.518819 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 10:08:40 crc kubenswrapper[4787]: I1001 10:08:40.518991 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vh5v6" Oct 01 10:08:40 crc kubenswrapper[4787]: I1001 10:08:40.519166 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 10:08:40 crc kubenswrapper[4787]: I1001 10:08:40.519321 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 01 10:08:40 crc kubenswrapper[4787]: I1001 10:08:40.577198 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr"] Oct 01 10:08:40 crc kubenswrapper[4787]: I1001 10:08:40.586437 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr\" (UID: \"c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr" Oct 01 10:08:40 crc kubenswrapper[4787]: I1001 10:08:40.586500 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr\" (UID: \"c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr" Oct 01 10:08:40 crc kubenswrapper[4787]: I1001 10:08:40.586532 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nv6j\" (UniqueName: \"kubernetes.io/projected/c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c-kube-api-access-8nv6j\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr\" (UID: \"c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr" Oct 01 10:08:40 crc kubenswrapper[4787]: I1001 10:08:40.586627 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr\" (UID: \"c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr" Oct 01 10:08:40 crc kubenswrapper[4787]: I1001 10:08:40.586670 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr\" (UID: \"c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr" Oct 01 10:08:40 crc kubenswrapper[4787]: I1001 10:08:40.688367 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr\" (UID: \"c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr" Oct 01 10:08:40 crc kubenswrapper[4787]: I1001 10:08:40.688424 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr\" (UID: \"c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr" Oct 01 10:08:40 crc kubenswrapper[4787]: I1001 10:08:40.688514 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr\" (UID: \"c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr" Oct 01 10:08:40 crc kubenswrapper[4787]: I1001 10:08:40.688556 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr\" (UID: \"c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr" Oct 01 10:08:40 crc kubenswrapper[4787]: I1001 10:08:40.688589 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nv6j\" (UniqueName: \"kubernetes.io/projected/c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c-kube-api-access-8nv6j\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr\" (UID: \"c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr" Oct 01 10:08:40 crc kubenswrapper[4787]: I1001 10:08:40.693430 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr\" (UID: \"c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr" Oct 01 10:08:40 crc kubenswrapper[4787]: I1001 10:08:40.693793 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr\" (UID: \"c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr" Oct 01 10:08:40 crc kubenswrapper[4787]: I1001 10:08:40.701285 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr\" (UID: \"c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr" Oct 01 10:08:40 crc kubenswrapper[4787]: I1001 10:08:40.708693 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr\" (UID: \"c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr" Oct 01 10:08:40 crc kubenswrapper[4787]: I1001 10:08:40.709907 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nv6j\" (UniqueName: \"kubernetes.io/projected/c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c-kube-api-access-8nv6j\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr\" (UID: \"c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr" Oct 01 10:08:40 crc kubenswrapper[4787]: I1001 10:08:40.886316 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr" Oct 01 10:08:41 crc kubenswrapper[4787]: I1001 10:08:41.250433 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:08:41 crc kubenswrapper[4787]: I1001 10:08:41.250495 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:08:41 crc kubenswrapper[4787]: I1001 10:08:41.407386 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr"] Oct 01 10:08:41 crc kubenswrapper[4787]: I1001 10:08:41.435237 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr" event={"ID":"c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c","Type":"ContainerStarted","Data":"56f019663262bf6ffa5e9111e49836db8ca9ae59ff474967eae184d61427d95d"} Oct 01 10:08:42 crc kubenswrapper[4787]: I1001 10:08:42.443261 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr" event={"ID":"c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c","Type":"ContainerStarted","Data":"4be59eb9074fa682c9f802e1d009f03e6a3a73a1e81a341734bdbc605f8deba7"} Oct 01 10:08:42 crc kubenswrapper[4787]: I1001 10:08:42.463946 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr" podStartSLOduration=1.9422212079999999 podStartE2EDuration="2.463927046s" podCreationTimestamp="2025-10-01 10:08:40 +0000 UTC" firstStartedPulling="2025-10-01 10:08:41.411165819 +0000 UTC m=+1953.526309976" lastFinishedPulling="2025-10-01 10:08:41.932871647 +0000 UTC m=+1954.048015814" observedRunningTime="2025-10-01 10:08:42.461421094 +0000 UTC m=+1954.576565271" watchObservedRunningTime="2025-10-01 10:08:42.463927046 +0000 UTC m=+1954.579071203" Oct 01 10:09:11 crc kubenswrapper[4787]: I1001 10:09:11.251030 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:09:11 crc kubenswrapper[4787]: I1001 10:09:11.252065 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:09:11 crc kubenswrapper[4787]: I1001 10:09:11.252148 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" Oct 01 10:09:11 crc kubenswrapper[4787]: I1001 10:09:11.253053 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0044a9c9898ad2d4aaf6c39b3a1b6ec6368111794783cf724347932ef58f6d3d"} pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 10:09:11 crc kubenswrapper[4787]: I1001 10:09:11.253151 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" containerID="cri-o://0044a9c9898ad2d4aaf6c39b3a1b6ec6368111794783cf724347932ef58f6d3d" gracePeriod=600 Oct 01 10:09:11 crc kubenswrapper[4787]: I1001 10:09:11.719930 4787 generic.go:334] "Generic (PLEG): container finished" podID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerID="0044a9c9898ad2d4aaf6c39b3a1b6ec6368111794783cf724347932ef58f6d3d" exitCode=0 Oct 01 10:09:11 crc kubenswrapper[4787]: I1001 10:09:11.719985 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" event={"ID":"fa51b883-5f3e-4141-9d4c-37704aac7718","Type":"ContainerDied","Data":"0044a9c9898ad2d4aaf6c39b3a1b6ec6368111794783cf724347932ef58f6d3d"} Oct 01 10:09:11 crc kubenswrapper[4787]: I1001 10:09:11.720509 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" event={"ID":"fa51b883-5f3e-4141-9d4c-37704aac7718","Type":"ContainerStarted","Data":"e05f5ad393ada7db54df5c58983c6c4e3e27cad1d8e65e9204e8b48c5e3ba7c1"} Oct 01 10:09:11 crc kubenswrapper[4787]: I1001 10:09:11.720535 4787 scope.go:117] "RemoveContainer" containerID="e01333312cf21546b537b606d139d2d34e67fa8885c08f9ceb1f4c790ae4fd3a" Oct 01 10:09:28 crc kubenswrapper[4787]: I1001 10:09:28.511302 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lcl9w"] Oct 01 10:09:28 crc kubenswrapper[4787]: I1001 10:09:28.513992 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lcl9w" Oct 01 10:09:28 crc kubenswrapper[4787]: I1001 10:09:28.536005 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lcl9w"] Oct 01 10:09:28 crc kubenswrapper[4787]: I1001 10:09:28.603247 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5klv5\" (UniqueName: \"kubernetes.io/projected/fc9fa5da-18cd-4d61-945a-f9278ec8d388-kube-api-access-5klv5\") pod \"community-operators-lcl9w\" (UID: \"fc9fa5da-18cd-4d61-945a-f9278ec8d388\") " pod="openshift-marketplace/community-operators-lcl9w" Oct 01 10:09:28 crc kubenswrapper[4787]: I1001 10:09:28.603326 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc9fa5da-18cd-4d61-945a-f9278ec8d388-utilities\") pod \"community-operators-lcl9w\" (UID: \"fc9fa5da-18cd-4d61-945a-f9278ec8d388\") " pod="openshift-marketplace/community-operators-lcl9w" Oct 01 10:09:28 crc kubenswrapper[4787]: I1001 10:09:28.603562 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc9fa5da-18cd-4d61-945a-f9278ec8d388-catalog-content\") pod \"community-operators-lcl9w\" (UID: \"fc9fa5da-18cd-4d61-945a-f9278ec8d388\") " pod="openshift-marketplace/community-operators-lcl9w" Oct 01 10:09:28 crc kubenswrapper[4787]: I1001 10:09:28.704250 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc9fa5da-18cd-4d61-945a-f9278ec8d388-catalog-content\") pod \"community-operators-lcl9w\" (UID: \"fc9fa5da-18cd-4d61-945a-f9278ec8d388\") " pod="openshift-marketplace/community-operators-lcl9w" Oct 01 10:09:28 crc kubenswrapper[4787]: I1001 10:09:28.704423 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5klv5\" (UniqueName: \"kubernetes.io/projected/fc9fa5da-18cd-4d61-945a-f9278ec8d388-kube-api-access-5klv5\") pod \"community-operators-lcl9w\" (UID: \"fc9fa5da-18cd-4d61-945a-f9278ec8d388\") " pod="openshift-marketplace/community-operators-lcl9w" Oct 01 10:09:28 crc kubenswrapper[4787]: I1001 10:09:28.704473 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc9fa5da-18cd-4d61-945a-f9278ec8d388-utilities\") pod \"community-operators-lcl9w\" (UID: \"fc9fa5da-18cd-4d61-945a-f9278ec8d388\") " pod="openshift-marketplace/community-operators-lcl9w" Oct 01 10:09:28 crc kubenswrapper[4787]: I1001 10:09:28.704889 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc9fa5da-18cd-4d61-945a-f9278ec8d388-catalog-content\") pod \"community-operators-lcl9w\" (UID: \"fc9fa5da-18cd-4d61-945a-f9278ec8d388\") " pod="openshift-marketplace/community-operators-lcl9w" Oct 01 10:09:28 crc kubenswrapper[4787]: I1001 10:09:28.704999 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc9fa5da-18cd-4d61-945a-f9278ec8d388-utilities\") pod \"community-operators-lcl9w\" (UID: \"fc9fa5da-18cd-4d61-945a-f9278ec8d388\") " pod="openshift-marketplace/community-operators-lcl9w" Oct 01 10:09:28 crc kubenswrapper[4787]: I1001 10:09:28.730034 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5klv5\" (UniqueName: \"kubernetes.io/projected/fc9fa5da-18cd-4d61-945a-f9278ec8d388-kube-api-access-5klv5\") pod \"community-operators-lcl9w\" (UID: \"fc9fa5da-18cd-4d61-945a-f9278ec8d388\") " pod="openshift-marketplace/community-operators-lcl9w" Oct 01 10:09:28 crc kubenswrapper[4787]: I1001 10:09:28.840607 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lcl9w" Oct 01 10:09:29 crc kubenswrapper[4787]: I1001 10:09:29.366970 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lcl9w"] Oct 01 10:09:29 crc kubenswrapper[4787]: W1001 10:09:29.370684 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc9fa5da_18cd_4d61_945a_f9278ec8d388.slice/crio-3a7fea6d4b151270eac7e470466a161e136f7e98b2efd6aa95ed976ba86d8fab WatchSource:0}: Error finding container 3a7fea6d4b151270eac7e470466a161e136f7e98b2efd6aa95ed976ba86d8fab: Status 404 returned error can't find the container with id 3a7fea6d4b151270eac7e470466a161e136f7e98b2efd6aa95ed976ba86d8fab Oct 01 10:09:29 crc kubenswrapper[4787]: E1001 10:09:29.715770 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc9fa5da_18cd_4d61_945a_f9278ec8d388.slice/crio-558606804caf0ff3b59d8f53ca1291e9e93a425af5914fa1ad09fa4b4bd59fb8.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc9fa5da_18cd_4d61_945a_f9278ec8d388.slice/crio-conmon-558606804caf0ff3b59d8f53ca1291e9e93a425af5914fa1ad09fa4b4bd59fb8.scope\": RecentStats: unable to find data in memory cache]" Oct 01 10:09:29 crc kubenswrapper[4787]: I1001 10:09:29.899698 4787 generic.go:334] "Generic (PLEG): container finished" podID="fc9fa5da-18cd-4d61-945a-f9278ec8d388" containerID="558606804caf0ff3b59d8f53ca1291e9e93a425af5914fa1ad09fa4b4bd59fb8" exitCode=0 Oct 01 10:09:29 crc kubenswrapper[4787]: I1001 10:09:29.899794 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lcl9w" event={"ID":"fc9fa5da-18cd-4d61-945a-f9278ec8d388","Type":"ContainerDied","Data":"558606804caf0ff3b59d8f53ca1291e9e93a425af5914fa1ad09fa4b4bd59fb8"} Oct 01 10:09:29 crc kubenswrapper[4787]: I1001 10:09:29.900529 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lcl9w" event={"ID":"fc9fa5da-18cd-4d61-945a-f9278ec8d388","Type":"ContainerStarted","Data":"3a7fea6d4b151270eac7e470466a161e136f7e98b2efd6aa95ed976ba86d8fab"} Oct 01 10:09:30 crc kubenswrapper[4787]: I1001 10:09:30.913708 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lcl9w" event={"ID":"fc9fa5da-18cd-4d61-945a-f9278ec8d388","Type":"ContainerStarted","Data":"73256fb12f7ca06a57155f7d3f83f25cb82307345ff66eeb029f67662b8da86f"} Oct 01 10:09:31 crc kubenswrapper[4787]: I1001 10:09:31.924960 4787 generic.go:334] "Generic (PLEG): container finished" podID="fc9fa5da-18cd-4d61-945a-f9278ec8d388" containerID="73256fb12f7ca06a57155f7d3f83f25cb82307345ff66eeb029f67662b8da86f" exitCode=0 Oct 01 10:09:31 crc kubenswrapper[4787]: I1001 10:09:31.925098 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lcl9w" event={"ID":"fc9fa5da-18cd-4d61-945a-f9278ec8d388","Type":"ContainerDied","Data":"73256fb12f7ca06a57155f7d3f83f25cb82307345ff66eeb029f67662b8da86f"} Oct 01 10:09:32 crc kubenswrapper[4787]: I1001 10:09:32.935461 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lcl9w" event={"ID":"fc9fa5da-18cd-4d61-945a-f9278ec8d388","Type":"ContainerStarted","Data":"2eb251d80faba47585ae35b8df718997a7ed8a646b588d4fbdd00cf9e2419f95"} Oct 01 10:09:32 crc kubenswrapper[4787]: I1001 10:09:32.962029 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lcl9w" podStartSLOduration=2.545802045 podStartE2EDuration="4.962010981s" podCreationTimestamp="2025-10-01 10:09:28 +0000 UTC" firstStartedPulling="2025-10-01 10:09:29.902464496 +0000 UTC m=+2002.017608653" lastFinishedPulling="2025-10-01 10:09:32.318673412 +0000 UTC m=+2004.433817589" observedRunningTime="2025-10-01 10:09:32.950574168 +0000 UTC m=+2005.065718315" watchObservedRunningTime="2025-10-01 10:09:32.962010981 +0000 UTC m=+2005.077155138" Oct 01 10:09:38 crc kubenswrapper[4787]: I1001 10:09:38.851269 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lcl9w" Oct 01 10:09:38 crc kubenswrapper[4787]: I1001 10:09:38.851969 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lcl9w" Oct 01 10:09:38 crc kubenswrapper[4787]: I1001 10:09:38.905086 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lcl9w" Oct 01 10:09:39 crc kubenswrapper[4787]: I1001 10:09:39.040607 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lcl9w" Oct 01 10:09:39 crc kubenswrapper[4787]: I1001 10:09:39.138411 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lcl9w"] Oct 01 10:09:41 crc kubenswrapper[4787]: I1001 10:09:41.015069 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lcl9w" podUID="fc9fa5da-18cd-4d61-945a-f9278ec8d388" containerName="registry-server" containerID="cri-o://2eb251d80faba47585ae35b8df718997a7ed8a646b588d4fbdd00cf9e2419f95" gracePeriod=2 Oct 01 10:09:42 crc kubenswrapper[4787]: I1001 10:09:42.023738 4787 generic.go:334] "Generic (PLEG): container finished" podID="fc9fa5da-18cd-4d61-945a-f9278ec8d388" containerID="2eb251d80faba47585ae35b8df718997a7ed8a646b588d4fbdd00cf9e2419f95" exitCode=0 Oct 01 10:09:42 crc kubenswrapper[4787]: I1001 10:09:42.023846 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lcl9w" event={"ID":"fc9fa5da-18cd-4d61-945a-f9278ec8d388","Type":"ContainerDied","Data":"2eb251d80faba47585ae35b8df718997a7ed8a646b588d4fbdd00cf9e2419f95"} Oct 01 10:09:42 crc kubenswrapper[4787]: I1001 10:09:42.024379 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lcl9w" event={"ID":"fc9fa5da-18cd-4d61-945a-f9278ec8d388","Type":"ContainerDied","Data":"3a7fea6d4b151270eac7e470466a161e136f7e98b2efd6aa95ed976ba86d8fab"} Oct 01 10:09:42 crc kubenswrapper[4787]: I1001 10:09:42.024400 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a7fea6d4b151270eac7e470466a161e136f7e98b2efd6aa95ed976ba86d8fab" Oct 01 10:09:42 crc kubenswrapper[4787]: I1001 10:09:42.065952 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lcl9w" Oct 01 10:09:42 crc kubenswrapper[4787]: I1001 10:09:42.166033 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc9fa5da-18cd-4d61-945a-f9278ec8d388-utilities\") pod \"fc9fa5da-18cd-4d61-945a-f9278ec8d388\" (UID: \"fc9fa5da-18cd-4d61-945a-f9278ec8d388\") " Oct 01 10:09:42 crc kubenswrapper[4787]: I1001 10:09:42.166179 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5klv5\" (UniqueName: \"kubernetes.io/projected/fc9fa5da-18cd-4d61-945a-f9278ec8d388-kube-api-access-5klv5\") pod \"fc9fa5da-18cd-4d61-945a-f9278ec8d388\" (UID: \"fc9fa5da-18cd-4d61-945a-f9278ec8d388\") " Oct 01 10:09:42 crc kubenswrapper[4787]: I1001 10:09:42.166266 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc9fa5da-18cd-4d61-945a-f9278ec8d388-catalog-content\") pod \"fc9fa5da-18cd-4d61-945a-f9278ec8d388\" (UID: \"fc9fa5da-18cd-4d61-945a-f9278ec8d388\") " Oct 01 10:09:42 crc kubenswrapper[4787]: I1001 10:09:42.167060 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc9fa5da-18cd-4d61-945a-f9278ec8d388-utilities" (OuterVolumeSpecName: "utilities") pod "fc9fa5da-18cd-4d61-945a-f9278ec8d388" (UID: "fc9fa5da-18cd-4d61-945a-f9278ec8d388"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:09:42 crc kubenswrapper[4787]: I1001 10:09:42.171956 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc9fa5da-18cd-4d61-945a-f9278ec8d388-kube-api-access-5klv5" (OuterVolumeSpecName: "kube-api-access-5klv5") pod "fc9fa5da-18cd-4d61-945a-f9278ec8d388" (UID: "fc9fa5da-18cd-4d61-945a-f9278ec8d388"). InnerVolumeSpecName "kube-api-access-5klv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:09:42 crc kubenswrapper[4787]: I1001 10:09:42.209577 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc9fa5da-18cd-4d61-945a-f9278ec8d388-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fc9fa5da-18cd-4d61-945a-f9278ec8d388" (UID: "fc9fa5da-18cd-4d61-945a-f9278ec8d388"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:09:42 crc kubenswrapper[4787]: I1001 10:09:42.268679 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5klv5\" (UniqueName: \"kubernetes.io/projected/fc9fa5da-18cd-4d61-945a-f9278ec8d388-kube-api-access-5klv5\") on node \"crc\" DevicePath \"\"" Oct 01 10:09:42 crc kubenswrapper[4787]: I1001 10:09:42.268725 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc9fa5da-18cd-4d61-945a-f9278ec8d388-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 10:09:42 crc kubenswrapper[4787]: I1001 10:09:42.268736 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc9fa5da-18cd-4d61-945a-f9278ec8d388-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 10:09:43 crc kubenswrapper[4787]: I1001 10:09:43.033634 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lcl9w" Oct 01 10:09:43 crc kubenswrapper[4787]: I1001 10:09:43.060709 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lcl9w"] Oct 01 10:09:43 crc kubenswrapper[4787]: I1001 10:09:43.068453 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lcl9w"] Oct 01 10:09:44 crc kubenswrapper[4787]: I1001 10:09:44.537478 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc9fa5da-18cd-4d61-945a-f9278ec8d388" path="/var/lib/kubelet/pods/fc9fa5da-18cd-4d61-945a-f9278ec8d388/volumes" Oct 01 10:09:44 crc kubenswrapper[4787]: I1001 10:09:44.549241 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-q265g"] Oct 01 10:09:44 crc kubenswrapper[4787]: E1001 10:09:44.549669 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc9fa5da-18cd-4d61-945a-f9278ec8d388" containerName="registry-server" Oct 01 10:09:44 crc kubenswrapper[4787]: I1001 10:09:44.549685 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc9fa5da-18cd-4d61-945a-f9278ec8d388" containerName="registry-server" Oct 01 10:09:44 crc kubenswrapper[4787]: E1001 10:09:44.549699 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc9fa5da-18cd-4d61-945a-f9278ec8d388" containerName="extract-utilities" Oct 01 10:09:44 crc kubenswrapper[4787]: I1001 10:09:44.549705 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc9fa5da-18cd-4d61-945a-f9278ec8d388" containerName="extract-utilities" Oct 01 10:09:44 crc kubenswrapper[4787]: E1001 10:09:44.549738 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc9fa5da-18cd-4d61-945a-f9278ec8d388" containerName="extract-content" Oct 01 10:09:44 crc kubenswrapper[4787]: I1001 10:09:44.549769 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc9fa5da-18cd-4d61-945a-f9278ec8d388" containerName="extract-content" Oct 01 10:09:44 crc kubenswrapper[4787]: I1001 10:09:44.549982 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc9fa5da-18cd-4d61-945a-f9278ec8d388" containerName="registry-server" Oct 01 10:09:44 crc kubenswrapper[4787]: I1001 10:09:44.552164 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q265g" Oct 01 10:09:44 crc kubenswrapper[4787]: I1001 10:09:44.578882 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q265g"] Oct 01 10:09:44 crc kubenswrapper[4787]: I1001 10:09:44.610776 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc633d34-26d5-4015-8800-b5568f588777-utilities\") pod \"redhat-operators-q265g\" (UID: \"bc633d34-26d5-4015-8800-b5568f588777\") " pod="openshift-marketplace/redhat-operators-q265g" Oct 01 10:09:44 crc kubenswrapper[4787]: I1001 10:09:44.611415 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc633d34-26d5-4015-8800-b5568f588777-catalog-content\") pod \"redhat-operators-q265g\" (UID: \"bc633d34-26d5-4015-8800-b5568f588777\") " pod="openshift-marketplace/redhat-operators-q265g" Oct 01 10:09:44 crc kubenswrapper[4787]: I1001 10:09:44.611709 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jbf8\" (UniqueName: \"kubernetes.io/projected/bc633d34-26d5-4015-8800-b5568f588777-kube-api-access-9jbf8\") pod \"redhat-operators-q265g\" (UID: \"bc633d34-26d5-4015-8800-b5568f588777\") " pod="openshift-marketplace/redhat-operators-q265g" Oct 01 10:09:44 crc kubenswrapper[4787]: I1001 10:09:44.713614 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc633d34-26d5-4015-8800-b5568f588777-utilities\") pod \"redhat-operators-q265g\" (UID: \"bc633d34-26d5-4015-8800-b5568f588777\") " pod="openshift-marketplace/redhat-operators-q265g" Oct 01 10:09:44 crc kubenswrapper[4787]: I1001 10:09:44.713738 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc633d34-26d5-4015-8800-b5568f588777-catalog-content\") pod \"redhat-operators-q265g\" (UID: \"bc633d34-26d5-4015-8800-b5568f588777\") " pod="openshift-marketplace/redhat-operators-q265g" Oct 01 10:09:44 crc kubenswrapper[4787]: I1001 10:09:44.713790 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jbf8\" (UniqueName: \"kubernetes.io/projected/bc633d34-26d5-4015-8800-b5568f588777-kube-api-access-9jbf8\") pod \"redhat-operators-q265g\" (UID: \"bc633d34-26d5-4015-8800-b5568f588777\") " pod="openshift-marketplace/redhat-operators-q265g" Oct 01 10:09:44 crc kubenswrapper[4787]: I1001 10:09:44.714670 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc633d34-26d5-4015-8800-b5568f588777-utilities\") pod \"redhat-operators-q265g\" (UID: \"bc633d34-26d5-4015-8800-b5568f588777\") " pod="openshift-marketplace/redhat-operators-q265g" Oct 01 10:09:44 crc kubenswrapper[4787]: I1001 10:09:44.714952 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc633d34-26d5-4015-8800-b5568f588777-catalog-content\") pod \"redhat-operators-q265g\" (UID: \"bc633d34-26d5-4015-8800-b5568f588777\") " pod="openshift-marketplace/redhat-operators-q265g" Oct 01 10:09:44 crc kubenswrapper[4787]: I1001 10:09:44.736969 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jbf8\" (UniqueName: \"kubernetes.io/projected/bc633d34-26d5-4015-8800-b5568f588777-kube-api-access-9jbf8\") pod \"redhat-operators-q265g\" (UID: \"bc633d34-26d5-4015-8800-b5568f588777\") " pod="openshift-marketplace/redhat-operators-q265g" Oct 01 10:09:44 crc kubenswrapper[4787]: I1001 10:09:44.880470 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q265g" Oct 01 10:09:45 crc kubenswrapper[4787]: I1001 10:09:45.341372 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q265g"] Oct 01 10:09:46 crc kubenswrapper[4787]: I1001 10:09:46.061898 4787 generic.go:334] "Generic (PLEG): container finished" podID="bc633d34-26d5-4015-8800-b5568f588777" containerID="f84d91500d976bedd40dc64bfa0d1a54abe2d02cf354b3654462ea221b0bd74c" exitCode=0 Oct 01 10:09:46 crc kubenswrapper[4787]: I1001 10:09:46.061971 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q265g" event={"ID":"bc633d34-26d5-4015-8800-b5568f588777","Type":"ContainerDied","Data":"f84d91500d976bedd40dc64bfa0d1a54abe2d02cf354b3654462ea221b0bd74c"} Oct 01 10:09:46 crc kubenswrapper[4787]: I1001 10:09:46.063226 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q265g" event={"ID":"bc633d34-26d5-4015-8800-b5568f588777","Type":"ContainerStarted","Data":"346d81fe6f681d914bfb751d5ca70382f8b50d1c1f7cf898bf2ef767abbb5500"} Oct 01 10:09:47 crc kubenswrapper[4787]: I1001 10:09:47.074874 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q265g" event={"ID":"bc633d34-26d5-4015-8800-b5568f588777","Type":"ContainerStarted","Data":"e2f8b2eb6e012e17eb9bb4731e479b2d2e73fe15cd35f589392409adc72b4a01"} Oct 01 10:09:48 crc kubenswrapper[4787]: I1001 10:09:48.084994 4787 generic.go:334] "Generic (PLEG): container finished" podID="bc633d34-26d5-4015-8800-b5568f588777" containerID="e2f8b2eb6e012e17eb9bb4731e479b2d2e73fe15cd35f589392409adc72b4a01" exitCode=0 Oct 01 10:09:48 crc kubenswrapper[4787]: I1001 10:09:48.085182 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q265g" event={"ID":"bc633d34-26d5-4015-8800-b5568f588777","Type":"ContainerDied","Data":"e2f8b2eb6e012e17eb9bb4731e479b2d2e73fe15cd35f589392409adc72b4a01"} Oct 01 10:09:49 crc kubenswrapper[4787]: I1001 10:09:49.096596 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q265g" event={"ID":"bc633d34-26d5-4015-8800-b5568f588777","Type":"ContainerStarted","Data":"634d3763d98e2152bf37cf8037c046d7743f937916200e537d1beb6051ea9c9d"} Oct 01 10:09:49 crc kubenswrapper[4787]: I1001 10:09:49.126695 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-q265g" podStartSLOduration=2.581781328 podStartE2EDuration="5.126678681s" podCreationTimestamp="2025-10-01 10:09:44 +0000 UTC" firstStartedPulling="2025-10-01 10:09:46.064368048 +0000 UTC m=+2018.179512205" lastFinishedPulling="2025-10-01 10:09:48.609265401 +0000 UTC m=+2020.724409558" observedRunningTime="2025-10-01 10:09:49.125270927 +0000 UTC m=+2021.240415124" watchObservedRunningTime="2025-10-01 10:09:49.126678681 +0000 UTC m=+2021.241822838" Oct 01 10:09:54 crc kubenswrapper[4787]: I1001 10:09:54.881740 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-q265g" Oct 01 10:09:54 crc kubenswrapper[4787]: I1001 10:09:54.882211 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-q265g" Oct 01 10:09:54 crc kubenswrapper[4787]: I1001 10:09:54.930226 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-q265g" Oct 01 10:09:55 crc kubenswrapper[4787]: I1001 10:09:55.194048 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-q265g" Oct 01 10:09:55 crc kubenswrapper[4787]: I1001 10:09:55.239480 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q265g"] Oct 01 10:09:57 crc kubenswrapper[4787]: I1001 10:09:57.165053 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-q265g" podUID="bc633d34-26d5-4015-8800-b5568f588777" containerName="registry-server" containerID="cri-o://634d3763d98e2152bf37cf8037c046d7743f937916200e537d1beb6051ea9c9d" gracePeriod=2 Oct 01 10:09:57 crc kubenswrapper[4787]: I1001 10:09:57.644791 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q265g" Oct 01 10:09:57 crc kubenswrapper[4787]: I1001 10:09:57.770422 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc633d34-26d5-4015-8800-b5568f588777-utilities\") pod \"bc633d34-26d5-4015-8800-b5568f588777\" (UID: \"bc633d34-26d5-4015-8800-b5568f588777\") " Oct 01 10:09:57 crc kubenswrapper[4787]: I1001 10:09:57.772275 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc633d34-26d5-4015-8800-b5568f588777-utilities" (OuterVolumeSpecName: "utilities") pod "bc633d34-26d5-4015-8800-b5568f588777" (UID: "bc633d34-26d5-4015-8800-b5568f588777"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:09:57 crc kubenswrapper[4787]: I1001 10:09:57.772472 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jbf8\" (UniqueName: \"kubernetes.io/projected/bc633d34-26d5-4015-8800-b5568f588777-kube-api-access-9jbf8\") pod \"bc633d34-26d5-4015-8800-b5568f588777\" (UID: \"bc633d34-26d5-4015-8800-b5568f588777\") " Oct 01 10:09:57 crc kubenswrapper[4787]: I1001 10:09:57.772635 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc633d34-26d5-4015-8800-b5568f588777-catalog-content\") pod \"bc633d34-26d5-4015-8800-b5568f588777\" (UID: \"bc633d34-26d5-4015-8800-b5568f588777\") " Oct 01 10:09:57 crc kubenswrapper[4787]: I1001 10:09:57.785727 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc633d34-26d5-4015-8800-b5568f588777-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 10:09:57 crc kubenswrapper[4787]: I1001 10:09:57.786545 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc633d34-26d5-4015-8800-b5568f588777-kube-api-access-9jbf8" (OuterVolumeSpecName: "kube-api-access-9jbf8") pod "bc633d34-26d5-4015-8800-b5568f588777" (UID: "bc633d34-26d5-4015-8800-b5568f588777"). InnerVolumeSpecName "kube-api-access-9jbf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:09:57 crc kubenswrapper[4787]: I1001 10:09:57.886748 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jbf8\" (UniqueName: \"kubernetes.io/projected/bc633d34-26d5-4015-8800-b5568f588777-kube-api-access-9jbf8\") on node \"crc\" DevicePath \"\"" Oct 01 10:09:58 crc kubenswrapper[4787]: I1001 10:09:58.060785 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc633d34-26d5-4015-8800-b5568f588777-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bc633d34-26d5-4015-8800-b5568f588777" (UID: "bc633d34-26d5-4015-8800-b5568f588777"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:09:58 crc kubenswrapper[4787]: I1001 10:09:58.090573 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc633d34-26d5-4015-8800-b5568f588777-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 10:09:58 crc kubenswrapper[4787]: I1001 10:09:58.178961 4787 generic.go:334] "Generic (PLEG): container finished" podID="bc633d34-26d5-4015-8800-b5568f588777" containerID="634d3763d98e2152bf37cf8037c046d7743f937916200e537d1beb6051ea9c9d" exitCode=0 Oct 01 10:09:58 crc kubenswrapper[4787]: I1001 10:09:58.179031 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q265g" Oct 01 10:09:58 crc kubenswrapper[4787]: I1001 10:09:58.179040 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q265g" event={"ID":"bc633d34-26d5-4015-8800-b5568f588777","Type":"ContainerDied","Data":"634d3763d98e2152bf37cf8037c046d7743f937916200e537d1beb6051ea9c9d"} Oct 01 10:09:58 crc kubenswrapper[4787]: I1001 10:09:58.179164 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q265g" event={"ID":"bc633d34-26d5-4015-8800-b5568f588777","Type":"ContainerDied","Data":"346d81fe6f681d914bfb751d5ca70382f8b50d1c1f7cf898bf2ef767abbb5500"} Oct 01 10:09:58 crc kubenswrapper[4787]: I1001 10:09:58.179190 4787 scope.go:117] "RemoveContainer" containerID="634d3763d98e2152bf37cf8037c046d7743f937916200e537d1beb6051ea9c9d" Oct 01 10:09:58 crc kubenswrapper[4787]: I1001 10:09:58.210049 4787 scope.go:117] "RemoveContainer" containerID="e2f8b2eb6e012e17eb9bb4731e479b2d2e73fe15cd35f589392409adc72b4a01" Oct 01 10:09:58 crc kubenswrapper[4787]: I1001 10:09:58.245510 4787 scope.go:117] "RemoveContainer" containerID="f84d91500d976bedd40dc64bfa0d1a54abe2d02cf354b3654462ea221b0bd74c" Oct 01 10:09:58 crc kubenswrapper[4787]: I1001 10:09:58.268839 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q265g"] Oct 01 10:09:58 crc kubenswrapper[4787]: I1001 10:09:58.279507 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-q265g"] Oct 01 10:09:58 crc kubenswrapper[4787]: I1001 10:09:58.304459 4787 scope.go:117] "RemoveContainer" containerID="634d3763d98e2152bf37cf8037c046d7743f937916200e537d1beb6051ea9c9d" Oct 01 10:09:58 crc kubenswrapper[4787]: E1001 10:09:58.305271 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"634d3763d98e2152bf37cf8037c046d7743f937916200e537d1beb6051ea9c9d\": container with ID starting with 634d3763d98e2152bf37cf8037c046d7743f937916200e537d1beb6051ea9c9d not found: ID does not exist" containerID="634d3763d98e2152bf37cf8037c046d7743f937916200e537d1beb6051ea9c9d" Oct 01 10:09:58 crc kubenswrapper[4787]: I1001 10:09:58.305348 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"634d3763d98e2152bf37cf8037c046d7743f937916200e537d1beb6051ea9c9d"} err="failed to get container status \"634d3763d98e2152bf37cf8037c046d7743f937916200e537d1beb6051ea9c9d\": rpc error: code = NotFound desc = could not find container \"634d3763d98e2152bf37cf8037c046d7743f937916200e537d1beb6051ea9c9d\": container with ID starting with 634d3763d98e2152bf37cf8037c046d7743f937916200e537d1beb6051ea9c9d not found: ID does not exist" Oct 01 10:09:58 crc kubenswrapper[4787]: I1001 10:09:58.305395 4787 scope.go:117] "RemoveContainer" containerID="e2f8b2eb6e012e17eb9bb4731e479b2d2e73fe15cd35f589392409adc72b4a01" Oct 01 10:09:58 crc kubenswrapper[4787]: E1001 10:09:58.305963 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2f8b2eb6e012e17eb9bb4731e479b2d2e73fe15cd35f589392409adc72b4a01\": container with ID starting with e2f8b2eb6e012e17eb9bb4731e479b2d2e73fe15cd35f589392409adc72b4a01 not found: ID does not exist" containerID="e2f8b2eb6e012e17eb9bb4731e479b2d2e73fe15cd35f589392409adc72b4a01" Oct 01 10:09:58 crc kubenswrapper[4787]: I1001 10:09:58.305997 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2f8b2eb6e012e17eb9bb4731e479b2d2e73fe15cd35f589392409adc72b4a01"} err="failed to get container status \"e2f8b2eb6e012e17eb9bb4731e479b2d2e73fe15cd35f589392409adc72b4a01\": rpc error: code = NotFound desc = could not find container \"e2f8b2eb6e012e17eb9bb4731e479b2d2e73fe15cd35f589392409adc72b4a01\": container with ID starting with e2f8b2eb6e012e17eb9bb4731e479b2d2e73fe15cd35f589392409adc72b4a01 not found: ID does not exist" Oct 01 10:09:58 crc kubenswrapper[4787]: I1001 10:09:58.306020 4787 scope.go:117] "RemoveContainer" containerID="f84d91500d976bedd40dc64bfa0d1a54abe2d02cf354b3654462ea221b0bd74c" Oct 01 10:09:58 crc kubenswrapper[4787]: E1001 10:09:58.306393 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f84d91500d976bedd40dc64bfa0d1a54abe2d02cf354b3654462ea221b0bd74c\": container with ID starting with f84d91500d976bedd40dc64bfa0d1a54abe2d02cf354b3654462ea221b0bd74c not found: ID does not exist" containerID="f84d91500d976bedd40dc64bfa0d1a54abe2d02cf354b3654462ea221b0bd74c" Oct 01 10:09:58 crc kubenswrapper[4787]: I1001 10:09:58.306434 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f84d91500d976bedd40dc64bfa0d1a54abe2d02cf354b3654462ea221b0bd74c"} err="failed to get container status \"f84d91500d976bedd40dc64bfa0d1a54abe2d02cf354b3654462ea221b0bd74c\": rpc error: code = NotFound desc = could not find container \"f84d91500d976bedd40dc64bfa0d1a54abe2d02cf354b3654462ea221b0bd74c\": container with ID starting with f84d91500d976bedd40dc64bfa0d1a54abe2d02cf354b3654462ea221b0bd74c not found: ID does not exist" Oct 01 10:09:58 crc kubenswrapper[4787]: I1001 10:09:58.535865 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc633d34-26d5-4015-8800-b5568f588777" path="/var/lib/kubelet/pods/bc633d34-26d5-4015-8800-b5568f588777/volumes" Oct 01 10:10:27 crc kubenswrapper[4787]: I1001 10:10:27.333703 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-l9j8h"] Oct 01 10:10:27 crc kubenswrapper[4787]: E1001 10:10:27.334703 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc633d34-26d5-4015-8800-b5568f588777" containerName="registry-server" Oct 01 10:10:27 crc kubenswrapper[4787]: I1001 10:10:27.334715 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc633d34-26d5-4015-8800-b5568f588777" containerName="registry-server" Oct 01 10:10:27 crc kubenswrapper[4787]: E1001 10:10:27.334729 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc633d34-26d5-4015-8800-b5568f588777" containerName="extract-utilities" Oct 01 10:10:27 crc kubenswrapper[4787]: I1001 10:10:27.334735 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc633d34-26d5-4015-8800-b5568f588777" containerName="extract-utilities" Oct 01 10:10:27 crc kubenswrapper[4787]: E1001 10:10:27.334755 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc633d34-26d5-4015-8800-b5568f588777" containerName="extract-content" Oct 01 10:10:27 crc kubenswrapper[4787]: I1001 10:10:27.334763 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc633d34-26d5-4015-8800-b5568f588777" containerName="extract-content" Oct 01 10:10:27 crc kubenswrapper[4787]: I1001 10:10:27.334961 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc633d34-26d5-4015-8800-b5568f588777" containerName="registry-server" Oct 01 10:10:27 crc kubenswrapper[4787]: I1001 10:10:27.336423 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l9j8h" Oct 01 10:10:27 crc kubenswrapper[4787]: I1001 10:10:27.343417 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l9j8h"] Oct 01 10:10:27 crc kubenswrapper[4787]: I1001 10:10:27.360136 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd-catalog-content\") pod \"certified-operators-l9j8h\" (UID: \"99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd\") " pod="openshift-marketplace/certified-operators-l9j8h" Oct 01 10:10:27 crc kubenswrapper[4787]: I1001 10:10:27.360237 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd-utilities\") pod \"certified-operators-l9j8h\" (UID: \"99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd\") " pod="openshift-marketplace/certified-operators-l9j8h" Oct 01 10:10:27 crc kubenswrapper[4787]: I1001 10:10:27.360272 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t296m\" (UniqueName: \"kubernetes.io/projected/99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd-kube-api-access-t296m\") pod \"certified-operators-l9j8h\" (UID: \"99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd\") " pod="openshift-marketplace/certified-operators-l9j8h" Oct 01 10:10:27 crc kubenswrapper[4787]: I1001 10:10:27.462516 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd-catalog-content\") pod \"certified-operators-l9j8h\" (UID: \"99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd\") " pod="openshift-marketplace/certified-operators-l9j8h" Oct 01 10:10:27 crc kubenswrapper[4787]: I1001 10:10:27.462712 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd-utilities\") pod \"certified-operators-l9j8h\" (UID: \"99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd\") " pod="openshift-marketplace/certified-operators-l9j8h" Oct 01 10:10:27 crc kubenswrapper[4787]: I1001 10:10:27.462767 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t296m\" (UniqueName: \"kubernetes.io/projected/99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd-kube-api-access-t296m\") pod \"certified-operators-l9j8h\" (UID: \"99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd\") " pod="openshift-marketplace/certified-operators-l9j8h" Oct 01 10:10:27 crc kubenswrapper[4787]: I1001 10:10:27.462988 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd-catalog-content\") pod \"certified-operators-l9j8h\" (UID: \"99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd\") " pod="openshift-marketplace/certified-operators-l9j8h" Oct 01 10:10:27 crc kubenswrapper[4787]: I1001 10:10:27.463565 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd-utilities\") pod \"certified-operators-l9j8h\" (UID: \"99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd\") " pod="openshift-marketplace/certified-operators-l9j8h" Oct 01 10:10:27 crc kubenswrapper[4787]: I1001 10:10:27.481666 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t296m\" (UniqueName: \"kubernetes.io/projected/99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd-kube-api-access-t296m\") pod \"certified-operators-l9j8h\" (UID: \"99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd\") " pod="openshift-marketplace/certified-operators-l9j8h" Oct 01 10:10:27 crc kubenswrapper[4787]: I1001 10:10:27.695190 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l9j8h" Oct 01 10:10:28 crc kubenswrapper[4787]: I1001 10:10:28.220668 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l9j8h"] Oct 01 10:10:28 crc kubenswrapper[4787]: I1001 10:10:28.454969 4787 generic.go:334] "Generic (PLEG): container finished" podID="99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd" containerID="626ef62b75c4d81fd45077aefc2b12fe4056f4c5fcd65fb86368b230ee932d4a" exitCode=0 Oct 01 10:10:28 crc kubenswrapper[4787]: I1001 10:10:28.455121 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l9j8h" event={"ID":"99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd","Type":"ContainerDied","Data":"626ef62b75c4d81fd45077aefc2b12fe4056f4c5fcd65fb86368b230ee932d4a"} Oct 01 10:10:28 crc kubenswrapper[4787]: I1001 10:10:28.455288 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l9j8h" event={"ID":"99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd","Type":"ContainerStarted","Data":"5ccfa7f8fc67de8913e35d63c98a6404737e9f3d2d18f262671e437d601991f9"} Oct 01 10:10:30 crc kubenswrapper[4787]: I1001 10:10:30.472166 4787 generic.go:334] "Generic (PLEG): container finished" podID="99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd" containerID="fe75468bb620f587f3648110dcc1af7fd734920d4850f21374cfb69d0ed32574" exitCode=0 Oct 01 10:10:30 crc kubenswrapper[4787]: I1001 10:10:30.472204 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l9j8h" event={"ID":"99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd","Type":"ContainerDied","Data":"fe75468bb620f587f3648110dcc1af7fd734920d4850f21374cfb69d0ed32574"} Oct 01 10:10:32 crc kubenswrapper[4787]: I1001 10:10:32.493621 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l9j8h" event={"ID":"99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd","Type":"ContainerStarted","Data":"b3ed7c88ea758117d04fc10037d7f76677c588670364de2ceee6527905a682db"} Oct 01 10:10:32 crc kubenswrapper[4787]: I1001 10:10:32.519111 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-l9j8h" podStartSLOduration=2.632942485 podStartE2EDuration="5.519092058s" podCreationTimestamp="2025-10-01 10:10:27 +0000 UTC" firstStartedPulling="2025-10-01 10:10:28.457001648 +0000 UTC m=+2060.572145805" lastFinishedPulling="2025-10-01 10:10:31.343151221 +0000 UTC m=+2063.458295378" observedRunningTime="2025-10-01 10:10:32.514016562 +0000 UTC m=+2064.629160739" watchObservedRunningTime="2025-10-01 10:10:32.519092058 +0000 UTC m=+2064.634236215" Oct 01 10:10:37 crc kubenswrapper[4787]: I1001 10:10:37.695462 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-l9j8h" Oct 01 10:10:37 crc kubenswrapper[4787]: I1001 10:10:37.696102 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-l9j8h" Oct 01 10:10:37 crc kubenswrapper[4787]: I1001 10:10:37.751278 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-l9j8h" Oct 01 10:10:38 crc kubenswrapper[4787]: I1001 10:10:38.599632 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-l9j8h" Oct 01 10:10:38 crc kubenswrapper[4787]: I1001 10:10:38.645718 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-l9j8h"] Oct 01 10:10:40 crc kubenswrapper[4787]: I1001 10:10:40.565929 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-l9j8h" podUID="99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd" containerName="registry-server" containerID="cri-o://b3ed7c88ea758117d04fc10037d7f76677c588670364de2ceee6527905a682db" gracePeriod=2 Oct 01 10:10:41 crc kubenswrapper[4787]: I1001 10:10:41.556892 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l9j8h" Oct 01 10:10:41 crc kubenswrapper[4787]: I1001 10:10:41.582397 4787 generic.go:334] "Generic (PLEG): container finished" podID="99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd" containerID="b3ed7c88ea758117d04fc10037d7f76677c588670364de2ceee6527905a682db" exitCode=0 Oct 01 10:10:41 crc kubenswrapper[4787]: I1001 10:10:41.582444 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l9j8h" event={"ID":"99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd","Type":"ContainerDied","Data":"b3ed7c88ea758117d04fc10037d7f76677c588670364de2ceee6527905a682db"} Oct 01 10:10:41 crc kubenswrapper[4787]: I1001 10:10:41.582465 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l9j8h" Oct 01 10:10:41 crc kubenswrapper[4787]: I1001 10:10:41.582484 4787 scope.go:117] "RemoveContainer" containerID="b3ed7c88ea758117d04fc10037d7f76677c588670364de2ceee6527905a682db" Oct 01 10:10:41 crc kubenswrapper[4787]: I1001 10:10:41.582472 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l9j8h" event={"ID":"99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd","Type":"ContainerDied","Data":"5ccfa7f8fc67de8913e35d63c98a6404737e9f3d2d18f262671e437d601991f9"} Oct 01 10:10:41 crc kubenswrapper[4787]: I1001 10:10:41.603909 4787 scope.go:117] "RemoveContainer" containerID="fe75468bb620f587f3648110dcc1af7fd734920d4850f21374cfb69d0ed32574" Oct 01 10:10:41 crc kubenswrapper[4787]: I1001 10:10:41.628934 4787 scope.go:117] "RemoveContainer" containerID="626ef62b75c4d81fd45077aefc2b12fe4056f4c5fcd65fb86368b230ee932d4a" Oct 01 10:10:41 crc kubenswrapper[4787]: I1001 10:10:41.666581 4787 scope.go:117] "RemoveContainer" containerID="b3ed7c88ea758117d04fc10037d7f76677c588670364de2ceee6527905a682db" Oct 01 10:10:41 crc kubenswrapper[4787]: E1001 10:10:41.667030 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3ed7c88ea758117d04fc10037d7f76677c588670364de2ceee6527905a682db\": container with ID starting with b3ed7c88ea758117d04fc10037d7f76677c588670364de2ceee6527905a682db not found: ID does not exist" containerID="b3ed7c88ea758117d04fc10037d7f76677c588670364de2ceee6527905a682db" Oct 01 10:10:41 crc kubenswrapper[4787]: I1001 10:10:41.667102 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3ed7c88ea758117d04fc10037d7f76677c588670364de2ceee6527905a682db"} err="failed to get container status \"b3ed7c88ea758117d04fc10037d7f76677c588670364de2ceee6527905a682db\": rpc error: code = NotFound desc = could not find container \"b3ed7c88ea758117d04fc10037d7f76677c588670364de2ceee6527905a682db\": container with ID starting with b3ed7c88ea758117d04fc10037d7f76677c588670364de2ceee6527905a682db not found: ID does not exist" Oct 01 10:10:41 crc kubenswrapper[4787]: I1001 10:10:41.667136 4787 scope.go:117] "RemoveContainer" containerID="fe75468bb620f587f3648110dcc1af7fd734920d4850f21374cfb69d0ed32574" Oct 01 10:10:41 crc kubenswrapper[4787]: E1001 10:10:41.667605 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe75468bb620f587f3648110dcc1af7fd734920d4850f21374cfb69d0ed32574\": container with ID starting with fe75468bb620f587f3648110dcc1af7fd734920d4850f21374cfb69d0ed32574 not found: ID does not exist" containerID="fe75468bb620f587f3648110dcc1af7fd734920d4850f21374cfb69d0ed32574" Oct 01 10:10:41 crc kubenswrapper[4787]: I1001 10:10:41.667635 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe75468bb620f587f3648110dcc1af7fd734920d4850f21374cfb69d0ed32574"} err="failed to get container status \"fe75468bb620f587f3648110dcc1af7fd734920d4850f21374cfb69d0ed32574\": rpc error: code = NotFound desc = could not find container \"fe75468bb620f587f3648110dcc1af7fd734920d4850f21374cfb69d0ed32574\": container with ID starting with fe75468bb620f587f3648110dcc1af7fd734920d4850f21374cfb69d0ed32574 not found: ID does not exist" Oct 01 10:10:41 crc kubenswrapper[4787]: I1001 10:10:41.667658 4787 scope.go:117] "RemoveContainer" containerID="626ef62b75c4d81fd45077aefc2b12fe4056f4c5fcd65fb86368b230ee932d4a" Oct 01 10:10:41 crc kubenswrapper[4787]: E1001 10:10:41.668171 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"626ef62b75c4d81fd45077aefc2b12fe4056f4c5fcd65fb86368b230ee932d4a\": container with ID starting with 626ef62b75c4d81fd45077aefc2b12fe4056f4c5fcd65fb86368b230ee932d4a not found: ID does not exist" containerID="626ef62b75c4d81fd45077aefc2b12fe4056f4c5fcd65fb86368b230ee932d4a" Oct 01 10:10:41 crc kubenswrapper[4787]: I1001 10:10:41.668214 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"626ef62b75c4d81fd45077aefc2b12fe4056f4c5fcd65fb86368b230ee932d4a"} err="failed to get container status \"626ef62b75c4d81fd45077aefc2b12fe4056f4c5fcd65fb86368b230ee932d4a\": rpc error: code = NotFound desc = could not find container \"626ef62b75c4d81fd45077aefc2b12fe4056f4c5fcd65fb86368b230ee932d4a\": container with ID starting with 626ef62b75c4d81fd45077aefc2b12fe4056f4c5fcd65fb86368b230ee932d4a not found: ID does not exist" Oct 01 10:10:41 crc kubenswrapper[4787]: I1001 10:10:41.720404 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd-catalog-content\") pod \"99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd\" (UID: \"99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd\") " Oct 01 10:10:41 crc kubenswrapper[4787]: I1001 10:10:41.720525 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd-utilities\") pod \"99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd\" (UID: \"99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd\") " Oct 01 10:10:41 crc kubenswrapper[4787]: I1001 10:10:41.720724 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t296m\" (UniqueName: \"kubernetes.io/projected/99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd-kube-api-access-t296m\") pod \"99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd\" (UID: \"99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd\") " Oct 01 10:10:41 crc kubenswrapper[4787]: I1001 10:10:41.723092 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd-utilities" (OuterVolumeSpecName: "utilities") pod "99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd" (UID: "99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:10:41 crc kubenswrapper[4787]: I1001 10:10:41.731444 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd-kube-api-access-t296m" (OuterVolumeSpecName: "kube-api-access-t296m") pod "99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd" (UID: "99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd"). InnerVolumeSpecName "kube-api-access-t296m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:10:41 crc kubenswrapper[4787]: I1001 10:10:41.770030 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd" (UID: "99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:10:41 crc kubenswrapper[4787]: I1001 10:10:41.823336 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 10:10:41 crc kubenswrapper[4787]: I1001 10:10:41.823374 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t296m\" (UniqueName: \"kubernetes.io/projected/99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd-kube-api-access-t296m\") on node \"crc\" DevicePath \"\"" Oct 01 10:10:41 crc kubenswrapper[4787]: I1001 10:10:41.823389 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 10:10:41 crc kubenswrapper[4787]: I1001 10:10:41.913708 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-l9j8h"] Oct 01 10:10:41 crc kubenswrapper[4787]: I1001 10:10:41.921744 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-l9j8h"] Oct 01 10:10:42 crc kubenswrapper[4787]: I1001 10:10:42.534615 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd" path="/var/lib/kubelet/pods/99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd/volumes" Oct 01 10:11:11 crc kubenswrapper[4787]: I1001 10:11:11.250979 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:11:11 crc kubenswrapper[4787]: I1001 10:11:11.251838 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:11:14 crc kubenswrapper[4787]: I1001 10:11:14.037904 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qzsqq"] Oct 01 10:11:14 crc kubenswrapper[4787]: E1001 10:11:14.038713 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd" containerName="registry-server" Oct 01 10:11:14 crc kubenswrapper[4787]: I1001 10:11:14.038729 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd" containerName="registry-server" Oct 01 10:11:14 crc kubenswrapper[4787]: E1001 10:11:14.038752 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd" containerName="extract-utilities" Oct 01 10:11:14 crc kubenswrapper[4787]: I1001 10:11:14.038761 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd" containerName="extract-utilities" Oct 01 10:11:14 crc kubenswrapper[4787]: E1001 10:11:14.038775 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd" containerName="extract-content" Oct 01 10:11:14 crc kubenswrapper[4787]: I1001 10:11:14.038784 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd" containerName="extract-content" Oct 01 10:11:14 crc kubenswrapper[4787]: I1001 10:11:14.039029 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="99b5e6a9-f4ae-42d6-bf77-b24e56e1d2fd" containerName="registry-server" Oct 01 10:11:14 crc kubenswrapper[4787]: I1001 10:11:14.040776 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qzsqq" Oct 01 10:11:14 crc kubenswrapper[4787]: I1001 10:11:14.050244 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qzsqq"] Oct 01 10:11:14 crc kubenswrapper[4787]: I1001 10:11:14.180024 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2dv6\" (UniqueName: \"kubernetes.io/projected/84d1f49c-9385-47fd-882d-827d073a9076-kube-api-access-b2dv6\") pod \"redhat-marketplace-qzsqq\" (UID: \"84d1f49c-9385-47fd-882d-827d073a9076\") " pod="openshift-marketplace/redhat-marketplace-qzsqq" Oct 01 10:11:14 crc kubenswrapper[4787]: I1001 10:11:14.180353 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84d1f49c-9385-47fd-882d-827d073a9076-catalog-content\") pod \"redhat-marketplace-qzsqq\" (UID: \"84d1f49c-9385-47fd-882d-827d073a9076\") " pod="openshift-marketplace/redhat-marketplace-qzsqq" Oct 01 10:11:14 crc kubenswrapper[4787]: I1001 10:11:14.180439 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84d1f49c-9385-47fd-882d-827d073a9076-utilities\") pod \"redhat-marketplace-qzsqq\" (UID: \"84d1f49c-9385-47fd-882d-827d073a9076\") " pod="openshift-marketplace/redhat-marketplace-qzsqq" Oct 01 10:11:14 crc kubenswrapper[4787]: I1001 10:11:14.282378 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84d1f49c-9385-47fd-882d-827d073a9076-catalog-content\") pod \"redhat-marketplace-qzsqq\" (UID: \"84d1f49c-9385-47fd-882d-827d073a9076\") " pod="openshift-marketplace/redhat-marketplace-qzsqq" Oct 01 10:11:14 crc kubenswrapper[4787]: I1001 10:11:14.282539 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84d1f49c-9385-47fd-882d-827d073a9076-utilities\") pod \"redhat-marketplace-qzsqq\" (UID: \"84d1f49c-9385-47fd-882d-827d073a9076\") " pod="openshift-marketplace/redhat-marketplace-qzsqq" Oct 01 10:11:14 crc kubenswrapper[4787]: I1001 10:11:14.282602 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2dv6\" (UniqueName: \"kubernetes.io/projected/84d1f49c-9385-47fd-882d-827d073a9076-kube-api-access-b2dv6\") pod \"redhat-marketplace-qzsqq\" (UID: \"84d1f49c-9385-47fd-882d-827d073a9076\") " pod="openshift-marketplace/redhat-marketplace-qzsqq" Oct 01 10:11:14 crc kubenswrapper[4787]: I1001 10:11:14.283021 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84d1f49c-9385-47fd-882d-827d073a9076-catalog-content\") pod \"redhat-marketplace-qzsqq\" (UID: \"84d1f49c-9385-47fd-882d-827d073a9076\") " pod="openshift-marketplace/redhat-marketplace-qzsqq" Oct 01 10:11:14 crc kubenswrapper[4787]: I1001 10:11:14.283566 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84d1f49c-9385-47fd-882d-827d073a9076-utilities\") pod \"redhat-marketplace-qzsqq\" (UID: \"84d1f49c-9385-47fd-882d-827d073a9076\") " pod="openshift-marketplace/redhat-marketplace-qzsqq" Oct 01 10:11:14 crc kubenswrapper[4787]: I1001 10:11:14.315062 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2dv6\" (UniqueName: \"kubernetes.io/projected/84d1f49c-9385-47fd-882d-827d073a9076-kube-api-access-b2dv6\") pod \"redhat-marketplace-qzsqq\" (UID: \"84d1f49c-9385-47fd-882d-827d073a9076\") " pod="openshift-marketplace/redhat-marketplace-qzsqq" Oct 01 10:11:14 crc kubenswrapper[4787]: I1001 10:11:14.373043 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qzsqq" Oct 01 10:11:14 crc kubenswrapper[4787]: I1001 10:11:14.814680 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qzsqq"] Oct 01 10:11:14 crc kubenswrapper[4787]: I1001 10:11:14.891144 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qzsqq" event={"ID":"84d1f49c-9385-47fd-882d-827d073a9076","Type":"ContainerStarted","Data":"6e7ef132e6e630279517f52dc9e69932e9d0536321ed16104fbaf15fd8c8b123"} Oct 01 10:11:15 crc kubenswrapper[4787]: I1001 10:11:15.907036 4787 generic.go:334] "Generic (PLEG): container finished" podID="84d1f49c-9385-47fd-882d-827d073a9076" containerID="a1c27d3ac78115300ea61d03378ba82a13a64f9120e20684a7c8e662f3db451f" exitCode=0 Oct 01 10:11:15 crc kubenswrapper[4787]: I1001 10:11:15.907114 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qzsqq" event={"ID":"84d1f49c-9385-47fd-882d-827d073a9076","Type":"ContainerDied","Data":"a1c27d3ac78115300ea61d03378ba82a13a64f9120e20684a7c8e662f3db451f"} Oct 01 10:11:16 crc kubenswrapper[4787]: I1001 10:11:16.916103 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qzsqq" event={"ID":"84d1f49c-9385-47fd-882d-827d073a9076","Type":"ContainerStarted","Data":"846b0d1dc25bda8b16036376d315ecde66742a7a958358b98b508f2823770d98"} Oct 01 10:11:17 crc kubenswrapper[4787]: I1001 10:11:17.925929 4787 generic.go:334] "Generic (PLEG): container finished" podID="84d1f49c-9385-47fd-882d-827d073a9076" containerID="846b0d1dc25bda8b16036376d315ecde66742a7a958358b98b508f2823770d98" exitCode=0 Oct 01 10:11:17 crc kubenswrapper[4787]: I1001 10:11:17.925978 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qzsqq" event={"ID":"84d1f49c-9385-47fd-882d-827d073a9076","Type":"ContainerDied","Data":"846b0d1dc25bda8b16036376d315ecde66742a7a958358b98b508f2823770d98"} Oct 01 10:11:18 crc kubenswrapper[4787]: I1001 10:11:18.937276 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qzsqq" event={"ID":"84d1f49c-9385-47fd-882d-827d073a9076","Type":"ContainerStarted","Data":"dee7af311dc8515811d3ad9333ef3a668e3be01ae67b20faf95e2c07b444e240"} Oct 01 10:11:18 crc kubenswrapper[4787]: I1001 10:11:18.958271 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qzsqq" podStartSLOduration=2.330567089 podStartE2EDuration="4.958252922s" podCreationTimestamp="2025-10-01 10:11:14 +0000 UTC" firstStartedPulling="2025-10-01 10:11:15.910353405 +0000 UTC m=+2108.025497602" lastFinishedPulling="2025-10-01 10:11:18.538039278 +0000 UTC m=+2110.653183435" observedRunningTime="2025-10-01 10:11:18.953738951 +0000 UTC m=+2111.068883118" watchObservedRunningTime="2025-10-01 10:11:18.958252922 +0000 UTC m=+2111.073397079" Oct 01 10:11:24 crc kubenswrapper[4787]: I1001 10:11:24.374177 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qzsqq" Oct 01 10:11:24 crc kubenswrapper[4787]: I1001 10:11:24.374772 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qzsqq" Oct 01 10:11:24 crc kubenswrapper[4787]: I1001 10:11:24.416623 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qzsqq" Oct 01 10:11:25 crc kubenswrapper[4787]: I1001 10:11:25.052150 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qzsqq" Oct 01 10:11:25 crc kubenswrapper[4787]: I1001 10:11:25.096199 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qzsqq"] Oct 01 10:11:27 crc kubenswrapper[4787]: I1001 10:11:27.027011 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qzsqq" podUID="84d1f49c-9385-47fd-882d-827d073a9076" containerName="registry-server" containerID="cri-o://dee7af311dc8515811d3ad9333ef3a668e3be01ae67b20faf95e2c07b444e240" gracePeriod=2 Oct 01 10:11:27 crc kubenswrapper[4787]: I1001 10:11:27.471020 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qzsqq" Oct 01 10:11:27 crc kubenswrapper[4787]: I1001 10:11:27.543634 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84d1f49c-9385-47fd-882d-827d073a9076-catalog-content\") pod \"84d1f49c-9385-47fd-882d-827d073a9076\" (UID: \"84d1f49c-9385-47fd-882d-827d073a9076\") " Oct 01 10:11:27 crc kubenswrapper[4787]: I1001 10:11:27.543832 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84d1f49c-9385-47fd-882d-827d073a9076-utilities\") pod \"84d1f49c-9385-47fd-882d-827d073a9076\" (UID: \"84d1f49c-9385-47fd-882d-827d073a9076\") " Oct 01 10:11:27 crc kubenswrapper[4787]: I1001 10:11:27.543875 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2dv6\" (UniqueName: \"kubernetes.io/projected/84d1f49c-9385-47fd-882d-827d073a9076-kube-api-access-b2dv6\") pod \"84d1f49c-9385-47fd-882d-827d073a9076\" (UID: \"84d1f49c-9385-47fd-882d-827d073a9076\") " Oct 01 10:11:27 crc kubenswrapper[4787]: I1001 10:11:27.550433 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84d1f49c-9385-47fd-882d-827d073a9076-kube-api-access-b2dv6" (OuterVolumeSpecName: "kube-api-access-b2dv6") pod "84d1f49c-9385-47fd-882d-827d073a9076" (UID: "84d1f49c-9385-47fd-882d-827d073a9076"). InnerVolumeSpecName "kube-api-access-b2dv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:11:27 crc kubenswrapper[4787]: I1001 10:11:27.551506 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84d1f49c-9385-47fd-882d-827d073a9076-utilities" (OuterVolumeSpecName: "utilities") pod "84d1f49c-9385-47fd-882d-827d073a9076" (UID: "84d1f49c-9385-47fd-882d-827d073a9076"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:11:27 crc kubenswrapper[4787]: I1001 10:11:27.568521 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84d1f49c-9385-47fd-882d-827d073a9076-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "84d1f49c-9385-47fd-882d-827d073a9076" (UID: "84d1f49c-9385-47fd-882d-827d073a9076"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:11:27 crc kubenswrapper[4787]: I1001 10:11:27.645924 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84d1f49c-9385-47fd-882d-827d073a9076-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 10:11:27 crc kubenswrapper[4787]: I1001 10:11:27.645951 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84d1f49c-9385-47fd-882d-827d073a9076-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 10:11:27 crc kubenswrapper[4787]: I1001 10:11:27.645960 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2dv6\" (UniqueName: \"kubernetes.io/projected/84d1f49c-9385-47fd-882d-827d073a9076-kube-api-access-b2dv6\") on node \"crc\" DevicePath \"\"" Oct 01 10:11:28 crc kubenswrapper[4787]: I1001 10:11:28.036744 4787 generic.go:334] "Generic (PLEG): container finished" podID="84d1f49c-9385-47fd-882d-827d073a9076" containerID="dee7af311dc8515811d3ad9333ef3a668e3be01ae67b20faf95e2c07b444e240" exitCode=0 Oct 01 10:11:28 crc kubenswrapper[4787]: I1001 10:11:28.036786 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qzsqq" event={"ID":"84d1f49c-9385-47fd-882d-827d073a9076","Type":"ContainerDied","Data":"dee7af311dc8515811d3ad9333ef3a668e3be01ae67b20faf95e2c07b444e240"} Oct 01 10:11:28 crc kubenswrapper[4787]: I1001 10:11:28.036812 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qzsqq" event={"ID":"84d1f49c-9385-47fd-882d-827d073a9076","Type":"ContainerDied","Data":"6e7ef132e6e630279517f52dc9e69932e9d0536321ed16104fbaf15fd8c8b123"} Oct 01 10:11:28 crc kubenswrapper[4787]: I1001 10:11:28.036828 4787 scope.go:117] "RemoveContainer" containerID="dee7af311dc8515811d3ad9333ef3a668e3be01ae67b20faf95e2c07b444e240" Oct 01 10:11:28 crc kubenswrapper[4787]: I1001 10:11:28.036855 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qzsqq" Oct 01 10:11:28 crc kubenswrapper[4787]: I1001 10:11:28.058995 4787 scope.go:117] "RemoveContainer" containerID="846b0d1dc25bda8b16036376d315ecde66742a7a958358b98b508f2823770d98" Oct 01 10:11:28 crc kubenswrapper[4787]: I1001 10:11:28.074313 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qzsqq"] Oct 01 10:11:28 crc kubenswrapper[4787]: I1001 10:11:28.082989 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qzsqq"] Oct 01 10:11:28 crc kubenswrapper[4787]: I1001 10:11:28.105060 4787 scope.go:117] "RemoveContainer" containerID="a1c27d3ac78115300ea61d03378ba82a13a64f9120e20684a7c8e662f3db451f" Oct 01 10:11:28 crc kubenswrapper[4787]: I1001 10:11:28.130188 4787 scope.go:117] "RemoveContainer" containerID="dee7af311dc8515811d3ad9333ef3a668e3be01ae67b20faf95e2c07b444e240" Oct 01 10:11:28 crc kubenswrapper[4787]: E1001 10:11:28.130759 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dee7af311dc8515811d3ad9333ef3a668e3be01ae67b20faf95e2c07b444e240\": container with ID starting with dee7af311dc8515811d3ad9333ef3a668e3be01ae67b20faf95e2c07b444e240 not found: ID does not exist" containerID="dee7af311dc8515811d3ad9333ef3a668e3be01ae67b20faf95e2c07b444e240" Oct 01 10:11:28 crc kubenswrapper[4787]: I1001 10:11:28.130858 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dee7af311dc8515811d3ad9333ef3a668e3be01ae67b20faf95e2c07b444e240"} err="failed to get container status \"dee7af311dc8515811d3ad9333ef3a668e3be01ae67b20faf95e2c07b444e240\": rpc error: code = NotFound desc = could not find container \"dee7af311dc8515811d3ad9333ef3a668e3be01ae67b20faf95e2c07b444e240\": container with ID starting with dee7af311dc8515811d3ad9333ef3a668e3be01ae67b20faf95e2c07b444e240 not found: ID does not exist" Oct 01 10:11:28 crc kubenswrapper[4787]: I1001 10:11:28.130918 4787 scope.go:117] "RemoveContainer" containerID="846b0d1dc25bda8b16036376d315ecde66742a7a958358b98b508f2823770d98" Oct 01 10:11:28 crc kubenswrapper[4787]: E1001 10:11:28.131352 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"846b0d1dc25bda8b16036376d315ecde66742a7a958358b98b508f2823770d98\": container with ID starting with 846b0d1dc25bda8b16036376d315ecde66742a7a958358b98b508f2823770d98 not found: ID does not exist" containerID="846b0d1dc25bda8b16036376d315ecde66742a7a958358b98b508f2823770d98" Oct 01 10:11:28 crc kubenswrapper[4787]: I1001 10:11:28.131383 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"846b0d1dc25bda8b16036376d315ecde66742a7a958358b98b508f2823770d98"} err="failed to get container status \"846b0d1dc25bda8b16036376d315ecde66742a7a958358b98b508f2823770d98\": rpc error: code = NotFound desc = could not find container \"846b0d1dc25bda8b16036376d315ecde66742a7a958358b98b508f2823770d98\": container with ID starting with 846b0d1dc25bda8b16036376d315ecde66742a7a958358b98b508f2823770d98 not found: ID does not exist" Oct 01 10:11:28 crc kubenswrapper[4787]: I1001 10:11:28.131405 4787 scope.go:117] "RemoveContainer" containerID="a1c27d3ac78115300ea61d03378ba82a13a64f9120e20684a7c8e662f3db451f" Oct 01 10:11:28 crc kubenswrapper[4787]: E1001 10:11:28.131618 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1c27d3ac78115300ea61d03378ba82a13a64f9120e20684a7c8e662f3db451f\": container with ID starting with a1c27d3ac78115300ea61d03378ba82a13a64f9120e20684a7c8e662f3db451f not found: ID does not exist" containerID="a1c27d3ac78115300ea61d03378ba82a13a64f9120e20684a7c8e662f3db451f" Oct 01 10:11:28 crc kubenswrapper[4787]: I1001 10:11:28.131644 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1c27d3ac78115300ea61d03378ba82a13a64f9120e20684a7c8e662f3db451f"} err="failed to get container status \"a1c27d3ac78115300ea61d03378ba82a13a64f9120e20684a7c8e662f3db451f\": rpc error: code = NotFound desc = could not find container \"a1c27d3ac78115300ea61d03378ba82a13a64f9120e20684a7c8e662f3db451f\": container with ID starting with a1c27d3ac78115300ea61d03378ba82a13a64f9120e20684a7c8e662f3db451f not found: ID does not exist" Oct 01 10:11:28 crc kubenswrapper[4787]: I1001 10:11:28.537459 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84d1f49c-9385-47fd-882d-827d073a9076" path="/var/lib/kubelet/pods/84d1f49c-9385-47fd-882d-827d073a9076/volumes" Oct 01 10:11:41 crc kubenswrapper[4787]: I1001 10:11:41.250442 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:11:41 crc kubenswrapper[4787]: I1001 10:11:41.251008 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:12:11 crc kubenswrapper[4787]: I1001 10:12:11.251171 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:12:11 crc kubenswrapper[4787]: I1001 10:12:11.251920 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:12:11 crc kubenswrapper[4787]: I1001 10:12:11.252027 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" Oct 01 10:12:11 crc kubenswrapper[4787]: I1001 10:12:11.252989 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e05f5ad393ada7db54df5c58983c6c4e3e27cad1d8e65e9204e8b48c5e3ba7c1"} pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 10:12:11 crc kubenswrapper[4787]: I1001 10:12:11.253059 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" containerID="cri-o://e05f5ad393ada7db54df5c58983c6c4e3e27cad1d8e65e9204e8b48c5e3ba7c1" gracePeriod=600 Oct 01 10:12:11 crc kubenswrapper[4787]: E1001 10:12:11.376067 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:12:11 crc kubenswrapper[4787]: I1001 10:12:11.449246 4787 generic.go:334] "Generic (PLEG): container finished" podID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerID="e05f5ad393ada7db54df5c58983c6c4e3e27cad1d8e65e9204e8b48c5e3ba7c1" exitCode=0 Oct 01 10:12:11 crc kubenswrapper[4787]: I1001 10:12:11.449292 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" event={"ID":"fa51b883-5f3e-4141-9d4c-37704aac7718","Type":"ContainerDied","Data":"e05f5ad393ada7db54df5c58983c6c4e3e27cad1d8e65e9204e8b48c5e3ba7c1"} Oct 01 10:12:11 crc kubenswrapper[4787]: I1001 10:12:11.449331 4787 scope.go:117] "RemoveContainer" containerID="0044a9c9898ad2d4aaf6c39b3a1b6ec6368111794783cf724347932ef58f6d3d" Oct 01 10:12:11 crc kubenswrapper[4787]: I1001 10:12:11.450067 4787 scope.go:117] "RemoveContainer" containerID="e05f5ad393ada7db54df5c58983c6c4e3e27cad1d8e65e9204e8b48c5e3ba7c1" Oct 01 10:12:11 crc kubenswrapper[4787]: E1001 10:12:11.450691 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:12:26 crc kubenswrapper[4787]: I1001 10:12:26.524056 4787 scope.go:117] "RemoveContainer" containerID="e05f5ad393ada7db54df5c58983c6c4e3e27cad1d8e65e9204e8b48c5e3ba7c1" Oct 01 10:12:26 crc kubenswrapper[4787]: E1001 10:12:26.524918 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:12:38 crc kubenswrapper[4787]: I1001 10:12:38.534508 4787 scope.go:117] "RemoveContainer" containerID="e05f5ad393ada7db54df5c58983c6c4e3e27cad1d8e65e9204e8b48c5e3ba7c1" Oct 01 10:12:38 crc kubenswrapper[4787]: E1001 10:12:38.535841 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:12:44 crc kubenswrapper[4787]: I1001 10:12:44.825659 4787 generic.go:334] "Generic (PLEG): container finished" podID="c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c" containerID="4be59eb9074fa682c9f802e1d009f03e6a3a73a1e81a341734bdbc605f8deba7" exitCode=0 Oct 01 10:12:44 crc kubenswrapper[4787]: I1001 10:12:44.825774 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr" event={"ID":"c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c","Type":"ContainerDied","Data":"4be59eb9074fa682c9f802e1d009f03e6a3a73a1e81a341734bdbc605f8deba7"} Oct 01 10:12:46 crc kubenswrapper[4787]: I1001 10:12:46.279233 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr" Oct 01 10:12:46 crc kubenswrapper[4787]: I1001 10:12:46.346301 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c-libvirt-secret-0\") pod \"c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c\" (UID: \"c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c\") " Oct 01 10:12:46 crc kubenswrapper[4787]: I1001 10:12:46.346421 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nv6j\" (UniqueName: \"kubernetes.io/projected/c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c-kube-api-access-8nv6j\") pod \"c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c\" (UID: \"c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c\") " Oct 01 10:12:46 crc kubenswrapper[4787]: I1001 10:12:46.346475 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c-libvirt-combined-ca-bundle\") pod \"c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c\" (UID: \"c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c\") " Oct 01 10:12:46 crc kubenswrapper[4787]: I1001 10:12:46.346566 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c-inventory\") pod \"c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c\" (UID: \"c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c\") " Oct 01 10:12:46 crc kubenswrapper[4787]: I1001 10:12:46.346612 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c-ssh-key\") pod \"c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c\" (UID: \"c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c\") " Oct 01 10:12:46 crc kubenswrapper[4787]: I1001 10:12:46.354857 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c" (UID: "c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:12:46 crc kubenswrapper[4787]: I1001 10:12:46.354930 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c-kube-api-access-8nv6j" (OuterVolumeSpecName: "kube-api-access-8nv6j") pod "c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c" (UID: "c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c"). InnerVolumeSpecName "kube-api-access-8nv6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:12:46 crc kubenswrapper[4787]: I1001 10:12:46.386164 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c-inventory" (OuterVolumeSpecName: "inventory") pod "c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c" (UID: "c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:12:46 crc kubenswrapper[4787]: I1001 10:12:46.388847 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c" (UID: "c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:12:46 crc kubenswrapper[4787]: I1001 10:12:46.389398 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c" (UID: "c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:12:46 crc kubenswrapper[4787]: I1001 10:12:46.448704 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 10:12:46 crc kubenswrapper[4787]: I1001 10:12:46.448731 4787 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 01 10:12:46 crc kubenswrapper[4787]: I1001 10:12:46.448742 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nv6j\" (UniqueName: \"kubernetes.io/projected/c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c-kube-api-access-8nv6j\") on node \"crc\" DevicePath \"\"" Oct 01 10:12:46 crc kubenswrapper[4787]: I1001 10:12:46.448751 4787 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 10:12:46 crc kubenswrapper[4787]: I1001 10:12:46.448760 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 10:12:46 crc kubenswrapper[4787]: I1001 10:12:46.845095 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr" event={"ID":"c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c","Type":"ContainerDied","Data":"56f019663262bf6ffa5e9111e49836db8ca9ae59ff474967eae184d61427d95d"} Oct 01 10:12:46 crc kubenswrapper[4787]: I1001 10:12:46.845136 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="56f019663262bf6ffa5e9111e49836db8ca9ae59ff474967eae184d61427d95d" Oct 01 10:12:46 crc kubenswrapper[4787]: I1001 10:12:46.845143 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr" Oct 01 10:12:46 crc kubenswrapper[4787]: I1001 10:12:46.948945 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-ltfkq"] Oct 01 10:12:46 crc kubenswrapper[4787]: E1001 10:12:46.949311 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84d1f49c-9385-47fd-882d-827d073a9076" containerName="registry-server" Oct 01 10:12:46 crc kubenswrapper[4787]: I1001 10:12:46.949329 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="84d1f49c-9385-47fd-882d-827d073a9076" containerName="registry-server" Oct 01 10:12:46 crc kubenswrapper[4787]: E1001 10:12:46.949346 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84d1f49c-9385-47fd-882d-827d073a9076" containerName="extract-utilities" Oct 01 10:12:46 crc kubenswrapper[4787]: I1001 10:12:46.949353 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="84d1f49c-9385-47fd-882d-827d073a9076" containerName="extract-utilities" Oct 01 10:12:46 crc kubenswrapper[4787]: E1001 10:12:46.949366 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 01 10:12:46 crc kubenswrapper[4787]: I1001 10:12:46.949373 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 01 10:12:46 crc kubenswrapper[4787]: E1001 10:12:46.949392 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84d1f49c-9385-47fd-882d-827d073a9076" containerName="extract-content" Oct 01 10:12:46 crc kubenswrapper[4787]: I1001 10:12:46.949399 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="84d1f49c-9385-47fd-882d-827d073a9076" containerName="extract-content" Oct 01 10:12:46 crc kubenswrapper[4787]: I1001 10:12:46.949578 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 01 10:12:46 crc kubenswrapper[4787]: I1001 10:12:46.949596 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="84d1f49c-9385-47fd-882d-827d073a9076" containerName="registry-server" Oct 01 10:12:46 crc kubenswrapper[4787]: I1001 10:12:46.950198 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ltfkq" Oct 01 10:12:46 crc kubenswrapper[4787]: I1001 10:12:46.952463 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vh5v6" Oct 01 10:12:46 crc kubenswrapper[4787]: I1001 10:12:46.952750 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 10:12:46 crc kubenswrapper[4787]: I1001 10:12:46.952872 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 01 10:12:46 crc kubenswrapper[4787]: I1001 10:12:46.952924 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 10:12:46 crc kubenswrapper[4787]: I1001 10:12:46.953023 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Oct 01 10:12:46 crc kubenswrapper[4787]: I1001 10:12:46.953440 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 10:12:46 crc kubenswrapper[4787]: I1001 10:12:46.954909 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 01 10:12:46 crc kubenswrapper[4787]: I1001 10:12:46.968366 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-ltfkq"] Oct 01 10:12:47 crc kubenswrapper[4787]: I1001 10:12:47.064111 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ltfkq\" (UID: \"4f22f28b-714f-4ab7-a90b-a400a500e810\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ltfkq" Oct 01 10:12:47 crc kubenswrapper[4787]: I1001 10:12:47.064174 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nq2ds\" (UniqueName: \"kubernetes.io/projected/4f22f28b-714f-4ab7-a90b-a400a500e810-kube-api-access-nq2ds\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ltfkq\" (UID: \"4f22f28b-714f-4ab7-a90b-a400a500e810\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ltfkq" Oct 01 10:12:47 crc kubenswrapper[4787]: I1001 10:12:47.064198 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ltfkq\" (UID: \"4f22f28b-714f-4ab7-a90b-a400a500e810\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ltfkq" Oct 01 10:12:47 crc kubenswrapper[4787]: I1001 10:12:47.064340 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ltfkq\" (UID: \"4f22f28b-714f-4ab7-a90b-a400a500e810\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ltfkq" Oct 01 10:12:47 crc kubenswrapper[4787]: I1001 10:12:47.064507 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ltfkq\" (UID: \"4f22f28b-714f-4ab7-a90b-a400a500e810\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ltfkq" Oct 01 10:12:47 crc kubenswrapper[4787]: I1001 10:12:47.064561 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ltfkq\" (UID: \"4f22f28b-714f-4ab7-a90b-a400a500e810\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ltfkq" Oct 01 10:12:47 crc kubenswrapper[4787]: I1001 10:12:47.064594 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ltfkq\" (UID: \"4f22f28b-714f-4ab7-a90b-a400a500e810\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ltfkq" Oct 01 10:12:47 crc kubenswrapper[4787]: I1001 10:12:47.064774 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ltfkq\" (UID: \"4f22f28b-714f-4ab7-a90b-a400a500e810\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ltfkq" Oct 01 10:12:47 crc kubenswrapper[4787]: I1001 10:12:47.064856 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4f22f28b-714f-4ab7-a90b-a400a500e810-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ltfkq\" (UID: \"4f22f28b-714f-4ab7-a90b-a400a500e810\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ltfkq" Oct 01 10:12:47 crc kubenswrapper[4787]: I1001 10:12:47.166925 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4f22f28b-714f-4ab7-a90b-a400a500e810-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ltfkq\" (UID: \"4f22f28b-714f-4ab7-a90b-a400a500e810\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ltfkq" Oct 01 10:12:47 crc kubenswrapper[4787]: I1001 10:12:47.167265 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ltfkq\" (UID: \"4f22f28b-714f-4ab7-a90b-a400a500e810\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ltfkq" Oct 01 10:12:47 crc kubenswrapper[4787]: I1001 10:12:47.167389 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nq2ds\" (UniqueName: \"kubernetes.io/projected/4f22f28b-714f-4ab7-a90b-a400a500e810-kube-api-access-nq2ds\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ltfkq\" (UID: \"4f22f28b-714f-4ab7-a90b-a400a500e810\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ltfkq" Oct 01 10:12:47 crc kubenswrapper[4787]: I1001 10:12:47.167422 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ltfkq\" (UID: \"4f22f28b-714f-4ab7-a90b-a400a500e810\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ltfkq" Oct 01 10:12:47 crc kubenswrapper[4787]: I1001 10:12:47.167489 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ltfkq\" (UID: \"4f22f28b-714f-4ab7-a90b-a400a500e810\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ltfkq" Oct 01 10:12:47 crc kubenswrapper[4787]: I1001 10:12:47.167581 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ltfkq\" (UID: \"4f22f28b-714f-4ab7-a90b-a400a500e810\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ltfkq" Oct 01 10:12:47 crc kubenswrapper[4787]: I1001 10:12:47.167616 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ltfkq\" (UID: \"4f22f28b-714f-4ab7-a90b-a400a500e810\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ltfkq" Oct 01 10:12:47 crc kubenswrapper[4787]: I1001 10:12:47.167639 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ltfkq\" (UID: \"4f22f28b-714f-4ab7-a90b-a400a500e810\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ltfkq" Oct 01 10:12:47 crc kubenswrapper[4787]: I1001 10:12:47.167720 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ltfkq\" (UID: \"4f22f28b-714f-4ab7-a90b-a400a500e810\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ltfkq" Oct 01 10:12:47 crc kubenswrapper[4787]: I1001 10:12:47.168286 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4f22f28b-714f-4ab7-a90b-a400a500e810-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ltfkq\" (UID: \"4f22f28b-714f-4ab7-a90b-a400a500e810\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ltfkq" Oct 01 10:12:47 crc kubenswrapper[4787]: I1001 10:12:47.171428 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ltfkq\" (UID: \"4f22f28b-714f-4ab7-a90b-a400a500e810\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ltfkq" Oct 01 10:12:47 crc kubenswrapper[4787]: I1001 10:12:47.171679 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ltfkq\" (UID: \"4f22f28b-714f-4ab7-a90b-a400a500e810\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ltfkq" Oct 01 10:12:47 crc kubenswrapper[4787]: I1001 10:12:47.172313 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ltfkq\" (UID: \"4f22f28b-714f-4ab7-a90b-a400a500e810\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ltfkq" Oct 01 10:12:47 crc kubenswrapper[4787]: I1001 10:12:47.172662 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ltfkq\" (UID: \"4f22f28b-714f-4ab7-a90b-a400a500e810\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ltfkq" Oct 01 10:12:47 crc kubenswrapper[4787]: I1001 10:12:47.172679 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ltfkq\" (UID: \"4f22f28b-714f-4ab7-a90b-a400a500e810\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ltfkq" Oct 01 10:12:47 crc kubenswrapper[4787]: I1001 10:12:47.172846 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ltfkq\" (UID: \"4f22f28b-714f-4ab7-a90b-a400a500e810\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ltfkq" Oct 01 10:12:47 crc kubenswrapper[4787]: I1001 10:12:47.175495 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ltfkq\" (UID: \"4f22f28b-714f-4ab7-a90b-a400a500e810\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ltfkq" Oct 01 10:12:47 crc kubenswrapper[4787]: I1001 10:12:47.187534 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nq2ds\" (UniqueName: \"kubernetes.io/projected/4f22f28b-714f-4ab7-a90b-a400a500e810-kube-api-access-nq2ds\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ltfkq\" (UID: \"4f22f28b-714f-4ab7-a90b-a400a500e810\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ltfkq" Oct 01 10:12:47 crc kubenswrapper[4787]: I1001 10:12:47.272761 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ltfkq" Oct 01 10:12:47 crc kubenswrapper[4787]: I1001 10:12:47.823322 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-ltfkq"] Oct 01 10:12:47 crc kubenswrapper[4787]: I1001 10:12:47.830106 4787 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 10:12:47 crc kubenswrapper[4787]: I1001 10:12:47.857486 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ltfkq" event={"ID":"4f22f28b-714f-4ab7-a90b-a400a500e810","Type":"ContainerStarted","Data":"b12a1ca9a6731e92967e628fa10e2e39db6322ff3e0de68401961b3518adc4b6"} Oct 01 10:12:48 crc kubenswrapper[4787]: I1001 10:12:48.869944 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ltfkq" event={"ID":"4f22f28b-714f-4ab7-a90b-a400a500e810","Type":"ContainerStarted","Data":"2f971b3ed67147d5770438eccf17c944334a116e129581ffd4c0569f76909d31"} Oct 01 10:12:48 crc kubenswrapper[4787]: I1001 10:12:48.895449 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ltfkq" podStartSLOduration=2.1297397670000002 podStartE2EDuration="2.895431604s" podCreationTimestamp="2025-10-01 10:12:46 +0000 UTC" firstStartedPulling="2025-10-01 10:12:47.829768417 +0000 UTC m=+2199.944912584" lastFinishedPulling="2025-10-01 10:12:48.595460264 +0000 UTC m=+2200.710604421" observedRunningTime="2025-10-01 10:12:48.892058093 +0000 UTC m=+2201.007202280" watchObservedRunningTime="2025-10-01 10:12:48.895431604 +0000 UTC m=+2201.010575771" Oct 01 10:12:49 crc kubenswrapper[4787]: I1001 10:12:49.524561 4787 scope.go:117] "RemoveContainer" containerID="e05f5ad393ada7db54df5c58983c6c4e3e27cad1d8e65e9204e8b48c5e3ba7c1" Oct 01 10:12:49 crc kubenswrapper[4787]: E1001 10:12:49.525032 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:13:04 crc kubenswrapper[4787]: I1001 10:13:04.524680 4787 scope.go:117] "RemoveContainer" containerID="e05f5ad393ada7db54df5c58983c6c4e3e27cad1d8e65e9204e8b48c5e3ba7c1" Oct 01 10:13:04 crc kubenswrapper[4787]: E1001 10:13:04.525547 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:13:15 crc kubenswrapper[4787]: I1001 10:13:15.524451 4787 scope.go:117] "RemoveContainer" containerID="e05f5ad393ada7db54df5c58983c6c4e3e27cad1d8e65e9204e8b48c5e3ba7c1" Oct 01 10:13:15 crc kubenswrapper[4787]: E1001 10:13:15.525423 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:13:26 crc kubenswrapper[4787]: I1001 10:13:26.523839 4787 scope.go:117] "RemoveContainer" containerID="e05f5ad393ada7db54df5c58983c6c4e3e27cad1d8e65e9204e8b48c5e3ba7c1" Oct 01 10:13:26 crc kubenswrapper[4787]: E1001 10:13:26.524794 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:13:39 crc kubenswrapper[4787]: I1001 10:13:39.523955 4787 scope.go:117] "RemoveContainer" containerID="e05f5ad393ada7db54df5c58983c6c4e3e27cad1d8e65e9204e8b48c5e3ba7c1" Oct 01 10:13:39 crc kubenswrapper[4787]: E1001 10:13:39.524746 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:13:52 crc kubenswrapper[4787]: I1001 10:13:52.524291 4787 scope.go:117] "RemoveContainer" containerID="e05f5ad393ada7db54df5c58983c6c4e3e27cad1d8e65e9204e8b48c5e3ba7c1" Oct 01 10:13:52 crc kubenswrapper[4787]: E1001 10:13:52.525462 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:14:03 crc kubenswrapper[4787]: I1001 10:14:03.525026 4787 scope.go:117] "RemoveContainer" containerID="e05f5ad393ada7db54df5c58983c6c4e3e27cad1d8e65e9204e8b48c5e3ba7c1" Oct 01 10:14:03 crc kubenswrapper[4787]: E1001 10:14:03.526276 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:14:16 crc kubenswrapper[4787]: I1001 10:14:16.523985 4787 scope.go:117] "RemoveContainer" containerID="e05f5ad393ada7db54df5c58983c6c4e3e27cad1d8e65e9204e8b48c5e3ba7c1" Oct 01 10:14:16 crc kubenswrapper[4787]: E1001 10:14:16.524883 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:14:30 crc kubenswrapper[4787]: I1001 10:14:30.523807 4787 scope.go:117] "RemoveContainer" containerID="e05f5ad393ada7db54df5c58983c6c4e3e27cad1d8e65e9204e8b48c5e3ba7c1" Oct 01 10:14:30 crc kubenswrapper[4787]: E1001 10:14:30.524663 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:14:42 crc kubenswrapper[4787]: I1001 10:14:42.523592 4787 scope.go:117] "RemoveContainer" containerID="e05f5ad393ada7db54df5c58983c6c4e3e27cad1d8e65e9204e8b48c5e3ba7c1" Oct 01 10:14:42 crc kubenswrapper[4787]: E1001 10:14:42.524406 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:14:53 crc kubenswrapper[4787]: I1001 10:14:53.523949 4787 scope.go:117] "RemoveContainer" containerID="e05f5ad393ada7db54df5c58983c6c4e3e27cad1d8e65e9204e8b48c5e3ba7c1" Oct 01 10:14:53 crc kubenswrapper[4787]: E1001 10:14:53.524651 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:15:00 crc kubenswrapper[4787]: I1001 10:15:00.161089 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321895-vmcsx"] Oct 01 10:15:00 crc kubenswrapper[4787]: I1001 10:15:00.164931 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321895-vmcsx" Oct 01 10:15:00 crc kubenswrapper[4787]: I1001 10:15:00.169855 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 01 10:15:00 crc kubenswrapper[4787]: I1001 10:15:00.169998 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 01 10:15:00 crc kubenswrapper[4787]: I1001 10:15:00.178771 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321895-vmcsx"] Oct 01 10:15:00 crc kubenswrapper[4787]: I1001 10:15:00.314068 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ff26d2a8-e6d7-477a-bf7e-a1c605a285e3-secret-volume\") pod \"collect-profiles-29321895-vmcsx\" (UID: \"ff26d2a8-e6d7-477a-bf7e-a1c605a285e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321895-vmcsx" Oct 01 10:15:00 crc kubenswrapper[4787]: I1001 10:15:00.314274 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ff26d2a8-e6d7-477a-bf7e-a1c605a285e3-config-volume\") pod \"collect-profiles-29321895-vmcsx\" (UID: \"ff26d2a8-e6d7-477a-bf7e-a1c605a285e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321895-vmcsx" Oct 01 10:15:00 crc kubenswrapper[4787]: I1001 10:15:00.314494 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmbzf\" (UniqueName: \"kubernetes.io/projected/ff26d2a8-e6d7-477a-bf7e-a1c605a285e3-kube-api-access-tmbzf\") pod \"collect-profiles-29321895-vmcsx\" (UID: \"ff26d2a8-e6d7-477a-bf7e-a1c605a285e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321895-vmcsx" Oct 01 10:15:00 crc kubenswrapper[4787]: I1001 10:15:00.417121 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ff26d2a8-e6d7-477a-bf7e-a1c605a285e3-config-volume\") pod \"collect-profiles-29321895-vmcsx\" (UID: \"ff26d2a8-e6d7-477a-bf7e-a1c605a285e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321895-vmcsx" Oct 01 10:15:00 crc kubenswrapper[4787]: I1001 10:15:00.417240 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmbzf\" (UniqueName: \"kubernetes.io/projected/ff26d2a8-e6d7-477a-bf7e-a1c605a285e3-kube-api-access-tmbzf\") pod \"collect-profiles-29321895-vmcsx\" (UID: \"ff26d2a8-e6d7-477a-bf7e-a1c605a285e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321895-vmcsx" Oct 01 10:15:00 crc kubenswrapper[4787]: I1001 10:15:00.417400 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ff26d2a8-e6d7-477a-bf7e-a1c605a285e3-secret-volume\") pod \"collect-profiles-29321895-vmcsx\" (UID: \"ff26d2a8-e6d7-477a-bf7e-a1c605a285e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321895-vmcsx" Oct 01 10:15:00 crc kubenswrapper[4787]: I1001 10:15:00.418555 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ff26d2a8-e6d7-477a-bf7e-a1c605a285e3-config-volume\") pod \"collect-profiles-29321895-vmcsx\" (UID: \"ff26d2a8-e6d7-477a-bf7e-a1c605a285e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321895-vmcsx" Oct 01 10:15:00 crc kubenswrapper[4787]: I1001 10:15:00.429878 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ff26d2a8-e6d7-477a-bf7e-a1c605a285e3-secret-volume\") pod \"collect-profiles-29321895-vmcsx\" (UID: \"ff26d2a8-e6d7-477a-bf7e-a1c605a285e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321895-vmcsx" Oct 01 10:15:00 crc kubenswrapper[4787]: I1001 10:15:00.437921 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmbzf\" (UniqueName: \"kubernetes.io/projected/ff26d2a8-e6d7-477a-bf7e-a1c605a285e3-kube-api-access-tmbzf\") pod \"collect-profiles-29321895-vmcsx\" (UID: \"ff26d2a8-e6d7-477a-bf7e-a1c605a285e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321895-vmcsx" Oct 01 10:15:00 crc kubenswrapper[4787]: I1001 10:15:00.501366 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321895-vmcsx" Oct 01 10:15:00 crc kubenswrapper[4787]: I1001 10:15:00.945187 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321895-vmcsx"] Oct 01 10:15:01 crc kubenswrapper[4787]: I1001 10:15:01.097282 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321895-vmcsx" event={"ID":"ff26d2a8-e6d7-477a-bf7e-a1c605a285e3","Type":"ContainerStarted","Data":"ebaa3c75dbe4d94c7d1a8329c17731b88bd861e387a025d3121224adad00f78f"} Oct 01 10:15:01 crc kubenswrapper[4787]: I1001 10:15:01.097326 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321895-vmcsx" event={"ID":"ff26d2a8-e6d7-477a-bf7e-a1c605a285e3","Type":"ContainerStarted","Data":"553dcc7c246c6c1ac18fd7326dc9485c828525d7e72522b99de7112f16aa7932"} Oct 01 10:15:01 crc kubenswrapper[4787]: I1001 10:15:01.118001 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29321895-vmcsx" podStartSLOduration=1.117982664 podStartE2EDuration="1.117982664s" podCreationTimestamp="2025-10-01 10:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 10:15:01.111708783 +0000 UTC m=+2333.226852940" watchObservedRunningTime="2025-10-01 10:15:01.117982664 +0000 UTC m=+2333.233126821" Oct 01 10:15:02 crc kubenswrapper[4787]: I1001 10:15:02.108803 4787 generic.go:334] "Generic (PLEG): container finished" podID="ff26d2a8-e6d7-477a-bf7e-a1c605a285e3" containerID="ebaa3c75dbe4d94c7d1a8329c17731b88bd861e387a025d3121224adad00f78f" exitCode=0 Oct 01 10:15:02 crc kubenswrapper[4787]: I1001 10:15:02.108920 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321895-vmcsx" event={"ID":"ff26d2a8-e6d7-477a-bf7e-a1c605a285e3","Type":"ContainerDied","Data":"ebaa3c75dbe4d94c7d1a8329c17731b88bd861e387a025d3121224adad00f78f"} Oct 01 10:15:03 crc kubenswrapper[4787]: I1001 10:15:03.460520 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321895-vmcsx" Oct 01 10:15:03 crc kubenswrapper[4787]: I1001 10:15:03.582062 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ff26d2a8-e6d7-477a-bf7e-a1c605a285e3-config-volume\") pod \"ff26d2a8-e6d7-477a-bf7e-a1c605a285e3\" (UID: \"ff26d2a8-e6d7-477a-bf7e-a1c605a285e3\") " Oct 01 10:15:03 crc kubenswrapper[4787]: I1001 10:15:03.582179 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ff26d2a8-e6d7-477a-bf7e-a1c605a285e3-secret-volume\") pod \"ff26d2a8-e6d7-477a-bf7e-a1c605a285e3\" (UID: \"ff26d2a8-e6d7-477a-bf7e-a1c605a285e3\") " Oct 01 10:15:03 crc kubenswrapper[4787]: I1001 10:15:03.582321 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmbzf\" (UniqueName: \"kubernetes.io/projected/ff26d2a8-e6d7-477a-bf7e-a1c605a285e3-kube-api-access-tmbzf\") pod \"ff26d2a8-e6d7-477a-bf7e-a1c605a285e3\" (UID: \"ff26d2a8-e6d7-477a-bf7e-a1c605a285e3\") " Oct 01 10:15:03 crc kubenswrapper[4787]: I1001 10:15:03.582858 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff26d2a8-e6d7-477a-bf7e-a1c605a285e3-config-volume" (OuterVolumeSpecName: "config-volume") pod "ff26d2a8-e6d7-477a-bf7e-a1c605a285e3" (UID: "ff26d2a8-e6d7-477a-bf7e-a1c605a285e3"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:15:03 crc kubenswrapper[4787]: I1001 10:15:03.583182 4787 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ff26d2a8-e6d7-477a-bf7e-a1c605a285e3-config-volume\") on node \"crc\" DevicePath \"\"" Oct 01 10:15:03 crc kubenswrapper[4787]: I1001 10:15:03.597787 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff26d2a8-e6d7-477a-bf7e-a1c605a285e3-kube-api-access-tmbzf" (OuterVolumeSpecName: "kube-api-access-tmbzf") pod "ff26d2a8-e6d7-477a-bf7e-a1c605a285e3" (UID: "ff26d2a8-e6d7-477a-bf7e-a1c605a285e3"). InnerVolumeSpecName "kube-api-access-tmbzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:15:03 crc kubenswrapper[4787]: I1001 10:15:03.604582 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff26d2a8-e6d7-477a-bf7e-a1c605a285e3-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ff26d2a8-e6d7-477a-bf7e-a1c605a285e3" (UID: "ff26d2a8-e6d7-477a-bf7e-a1c605a285e3"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:15:03 crc kubenswrapper[4787]: I1001 10:15:03.684740 4787 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ff26d2a8-e6d7-477a-bf7e-a1c605a285e3-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 01 10:15:03 crc kubenswrapper[4787]: I1001 10:15:03.684770 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmbzf\" (UniqueName: \"kubernetes.io/projected/ff26d2a8-e6d7-477a-bf7e-a1c605a285e3-kube-api-access-tmbzf\") on node \"crc\" DevicePath \"\"" Oct 01 10:15:04 crc kubenswrapper[4787]: I1001 10:15:04.125456 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321895-vmcsx" event={"ID":"ff26d2a8-e6d7-477a-bf7e-a1c605a285e3","Type":"ContainerDied","Data":"553dcc7c246c6c1ac18fd7326dc9485c828525d7e72522b99de7112f16aa7932"} Oct 01 10:15:04 crc kubenswrapper[4787]: I1001 10:15:04.125497 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="553dcc7c246c6c1ac18fd7326dc9485c828525d7e72522b99de7112f16aa7932" Oct 01 10:15:04 crc kubenswrapper[4787]: I1001 10:15:04.125554 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321895-vmcsx" Oct 01 10:15:04 crc kubenswrapper[4787]: I1001 10:15:04.545102 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321850-kw99x"] Oct 01 10:15:04 crc kubenswrapper[4787]: I1001 10:15:04.553058 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321850-kw99x"] Oct 01 10:15:06 crc kubenswrapper[4787]: I1001 10:15:06.524302 4787 scope.go:117] "RemoveContainer" containerID="e05f5ad393ada7db54df5c58983c6c4e3e27cad1d8e65e9204e8b48c5e3ba7c1" Oct 01 10:15:06 crc kubenswrapper[4787]: E1001 10:15:06.524860 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:15:06 crc kubenswrapper[4787]: I1001 10:15:06.534040 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8de4a645-2945-4261-b9be-ee3336e818df" path="/var/lib/kubelet/pods/8de4a645-2945-4261-b9be-ee3336e818df/volumes" Oct 01 10:15:20 crc kubenswrapper[4787]: I1001 10:15:20.524705 4787 scope.go:117] "RemoveContainer" containerID="e05f5ad393ada7db54df5c58983c6c4e3e27cad1d8e65e9204e8b48c5e3ba7c1" Oct 01 10:15:20 crc kubenswrapper[4787]: E1001 10:15:20.525490 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:15:31 crc kubenswrapper[4787]: I1001 10:15:31.524525 4787 scope.go:117] "RemoveContainer" containerID="e05f5ad393ada7db54df5c58983c6c4e3e27cad1d8e65e9204e8b48c5e3ba7c1" Oct 01 10:15:31 crc kubenswrapper[4787]: E1001 10:15:31.525752 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:15:32 crc kubenswrapper[4787]: I1001 10:15:32.575914 4787 scope.go:117] "RemoveContainer" containerID="73256fb12f7ca06a57155f7d3f83f25cb82307345ff66eeb029f67662b8da86f" Oct 01 10:15:32 crc kubenswrapper[4787]: I1001 10:15:32.603107 4787 scope.go:117] "RemoveContainer" containerID="2eb251d80faba47585ae35b8df718997a7ed8a646b588d4fbdd00cf9e2419f95" Oct 01 10:15:32 crc kubenswrapper[4787]: I1001 10:15:32.646220 4787 scope.go:117] "RemoveContainer" containerID="558606804caf0ff3b59d8f53ca1291e9e93a425af5914fa1ad09fa4b4bd59fb8" Oct 01 10:15:32 crc kubenswrapper[4787]: I1001 10:15:32.663700 4787 scope.go:117] "RemoveContainer" containerID="7c1e9231673196bcc39c382d8fed3c520fb5a11bde8abe81df02128bbaefbe48" Oct 01 10:15:43 crc kubenswrapper[4787]: I1001 10:15:43.524130 4787 scope.go:117] "RemoveContainer" containerID="e05f5ad393ada7db54df5c58983c6c4e3e27cad1d8e65e9204e8b48c5e3ba7c1" Oct 01 10:15:43 crc kubenswrapper[4787]: E1001 10:15:43.524795 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:15:55 crc kubenswrapper[4787]: I1001 10:15:55.523976 4787 scope.go:117] "RemoveContainer" containerID="e05f5ad393ada7db54df5c58983c6c4e3e27cad1d8e65e9204e8b48c5e3ba7c1" Oct 01 10:15:55 crc kubenswrapper[4787]: E1001 10:15:55.525541 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:16:05 crc kubenswrapper[4787]: I1001 10:16:05.777300 4787 generic.go:334] "Generic (PLEG): container finished" podID="4f22f28b-714f-4ab7-a90b-a400a500e810" containerID="2f971b3ed67147d5770438eccf17c944334a116e129581ffd4c0569f76909d31" exitCode=0 Oct 01 10:16:05 crc kubenswrapper[4787]: I1001 10:16:05.777458 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ltfkq" event={"ID":"4f22f28b-714f-4ab7-a90b-a400a500e810","Type":"ContainerDied","Data":"2f971b3ed67147d5770438eccf17c944334a116e129581ffd4c0569f76909d31"} Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.222719 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ltfkq" Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.317723 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-inventory\") pod \"4f22f28b-714f-4ab7-a90b-a400a500e810\" (UID: \"4f22f28b-714f-4ab7-a90b-a400a500e810\") " Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.317809 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-nova-cell1-compute-config-1\") pod \"4f22f28b-714f-4ab7-a90b-a400a500e810\" (UID: \"4f22f28b-714f-4ab7-a90b-a400a500e810\") " Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.317879 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-nova-migration-ssh-key-1\") pod \"4f22f28b-714f-4ab7-a90b-a400a500e810\" (UID: \"4f22f28b-714f-4ab7-a90b-a400a500e810\") " Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.317902 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-ssh-key\") pod \"4f22f28b-714f-4ab7-a90b-a400a500e810\" (UID: \"4f22f28b-714f-4ab7-a90b-a400a500e810\") " Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.317987 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nq2ds\" (UniqueName: \"kubernetes.io/projected/4f22f28b-714f-4ab7-a90b-a400a500e810-kube-api-access-nq2ds\") pod \"4f22f28b-714f-4ab7-a90b-a400a500e810\" (UID: \"4f22f28b-714f-4ab7-a90b-a400a500e810\") " Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.318041 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-nova-combined-ca-bundle\") pod \"4f22f28b-714f-4ab7-a90b-a400a500e810\" (UID: \"4f22f28b-714f-4ab7-a90b-a400a500e810\") " Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.318106 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-nova-migration-ssh-key-0\") pod \"4f22f28b-714f-4ab7-a90b-a400a500e810\" (UID: \"4f22f28b-714f-4ab7-a90b-a400a500e810\") " Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.318142 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4f22f28b-714f-4ab7-a90b-a400a500e810-nova-extra-config-0\") pod \"4f22f28b-714f-4ab7-a90b-a400a500e810\" (UID: \"4f22f28b-714f-4ab7-a90b-a400a500e810\") " Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.318181 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-nova-cell1-compute-config-0\") pod \"4f22f28b-714f-4ab7-a90b-a400a500e810\" (UID: \"4f22f28b-714f-4ab7-a90b-a400a500e810\") " Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.328697 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "4f22f28b-714f-4ab7-a90b-a400a500e810" (UID: "4f22f28b-714f-4ab7-a90b-a400a500e810"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.331665 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f22f28b-714f-4ab7-a90b-a400a500e810-kube-api-access-nq2ds" (OuterVolumeSpecName: "kube-api-access-nq2ds") pod "4f22f28b-714f-4ab7-a90b-a400a500e810" (UID: "4f22f28b-714f-4ab7-a90b-a400a500e810"). InnerVolumeSpecName "kube-api-access-nq2ds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.346205 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "4f22f28b-714f-4ab7-a90b-a400a500e810" (UID: "4f22f28b-714f-4ab7-a90b-a400a500e810"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.351246 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4f22f28b-714f-4ab7-a90b-a400a500e810" (UID: "4f22f28b-714f-4ab7-a90b-a400a500e810"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.351640 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f22f28b-714f-4ab7-a90b-a400a500e810-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "4f22f28b-714f-4ab7-a90b-a400a500e810" (UID: "4f22f28b-714f-4ab7-a90b-a400a500e810"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.352138 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-inventory" (OuterVolumeSpecName: "inventory") pod "4f22f28b-714f-4ab7-a90b-a400a500e810" (UID: "4f22f28b-714f-4ab7-a90b-a400a500e810"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.353861 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "4f22f28b-714f-4ab7-a90b-a400a500e810" (UID: "4f22f28b-714f-4ab7-a90b-a400a500e810"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.354671 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "4f22f28b-714f-4ab7-a90b-a400a500e810" (UID: "4f22f28b-714f-4ab7-a90b-a400a500e810"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.356375 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "4f22f28b-714f-4ab7-a90b-a400a500e810" (UID: "4f22f28b-714f-4ab7-a90b-a400a500e810"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.419990 4787 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.420040 4787 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.420052 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.420063 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nq2ds\" (UniqueName: \"kubernetes.io/projected/4f22f28b-714f-4ab7-a90b-a400a500e810-kube-api-access-nq2ds\") on node \"crc\" DevicePath \"\"" Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.420093 4787 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.420105 4787 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.420120 4787 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4f22f28b-714f-4ab7-a90b-a400a500e810-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.420132 4787 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.420144 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f22f28b-714f-4ab7-a90b-a400a500e810-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.523636 4787 scope.go:117] "RemoveContainer" containerID="e05f5ad393ada7db54df5c58983c6c4e3e27cad1d8e65e9204e8b48c5e3ba7c1" Oct 01 10:16:07 crc kubenswrapper[4787]: E1001 10:16:07.523851 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.797917 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ltfkq" Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.797892 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ltfkq" event={"ID":"4f22f28b-714f-4ab7-a90b-a400a500e810","Type":"ContainerDied","Data":"b12a1ca9a6731e92967e628fa10e2e39db6322ff3e0de68401961b3518adc4b6"} Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.798277 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b12a1ca9a6731e92967e628fa10e2e39db6322ff3e0de68401961b3518adc4b6" Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.905586 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9"] Oct 01 10:16:07 crc kubenswrapper[4787]: E1001 10:16:07.906010 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f22f28b-714f-4ab7-a90b-a400a500e810" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.906029 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f22f28b-714f-4ab7-a90b-a400a500e810" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 01 10:16:07 crc kubenswrapper[4787]: E1001 10:16:07.906041 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff26d2a8-e6d7-477a-bf7e-a1c605a285e3" containerName="collect-profiles" Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.906047 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff26d2a8-e6d7-477a-bf7e-a1c605a285e3" containerName="collect-profiles" Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.906258 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f22f28b-714f-4ab7-a90b-a400a500e810" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.906280 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff26d2a8-e6d7-477a-bf7e-a1c605a285e3" containerName="collect-profiles" Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.906958 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9" Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.909273 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.909352 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.909443 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.911590 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-vh5v6" Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.912012 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 10:16:07 crc kubenswrapper[4787]: I1001 10:16:07.924504 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9"] Oct 01 10:16:08 crc kubenswrapper[4787]: I1001 10:16:08.038529 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9\" (UID: \"56b1c51b-bdcd-4c55-a00d-105ab4bc04b0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9" Oct 01 10:16:08 crc kubenswrapper[4787]: I1001 10:16:08.038598 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9\" (UID: \"56b1c51b-bdcd-4c55-a00d-105ab4bc04b0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9" Oct 01 10:16:08 crc kubenswrapper[4787]: I1001 10:16:08.038701 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9\" (UID: \"56b1c51b-bdcd-4c55-a00d-105ab4bc04b0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9" Oct 01 10:16:08 crc kubenswrapper[4787]: I1001 10:16:08.038755 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9\" (UID: \"56b1c51b-bdcd-4c55-a00d-105ab4bc04b0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9" Oct 01 10:16:08 crc kubenswrapper[4787]: I1001 10:16:08.038802 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9\" (UID: \"56b1c51b-bdcd-4c55-a00d-105ab4bc04b0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9" Oct 01 10:16:08 crc kubenswrapper[4787]: I1001 10:16:08.038903 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42g4q\" (UniqueName: \"kubernetes.io/projected/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-kube-api-access-42g4q\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9\" (UID: \"56b1c51b-bdcd-4c55-a00d-105ab4bc04b0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9" Oct 01 10:16:08 crc kubenswrapper[4787]: I1001 10:16:08.038950 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9\" (UID: \"56b1c51b-bdcd-4c55-a00d-105ab4bc04b0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9" Oct 01 10:16:08 crc kubenswrapper[4787]: I1001 10:16:08.140851 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42g4q\" (UniqueName: \"kubernetes.io/projected/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-kube-api-access-42g4q\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9\" (UID: \"56b1c51b-bdcd-4c55-a00d-105ab4bc04b0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9" Oct 01 10:16:08 crc kubenswrapper[4787]: I1001 10:16:08.140911 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9\" (UID: \"56b1c51b-bdcd-4c55-a00d-105ab4bc04b0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9" Oct 01 10:16:08 crc kubenswrapper[4787]: I1001 10:16:08.140982 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9\" (UID: \"56b1c51b-bdcd-4c55-a00d-105ab4bc04b0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9" Oct 01 10:16:08 crc kubenswrapper[4787]: I1001 10:16:08.141017 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9\" (UID: \"56b1c51b-bdcd-4c55-a00d-105ab4bc04b0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9" Oct 01 10:16:08 crc kubenswrapper[4787]: I1001 10:16:08.141053 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9\" (UID: \"56b1c51b-bdcd-4c55-a00d-105ab4bc04b0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9" Oct 01 10:16:08 crc kubenswrapper[4787]: I1001 10:16:08.141099 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9\" (UID: \"56b1c51b-bdcd-4c55-a00d-105ab4bc04b0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9" Oct 01 10:16:08 crc kubenswrapper[4787]: I1001 10:16:08.141122 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9\" (UID: \"56b1c51b-bdcd-4c55-a00d-105ab4bc04b0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9" Oct 01 10:16:08 crc kubenswrapper[4787]: I1001 10:16:08.146673 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9\" (UID: \"56b1c51b-bdcd-4c55-a00d-105ab4bc04b0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9" Oct 01 10:16:08 crc kubenswrapper[4787]: I1001 10:16:08.146727 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9\" (UID: \"56b1c51b-bdcd-4c55-a00d-105ab4bc04b0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9" Oct 01 10:16:08 crc kubenswrapper[4787]: I1001 10:16:08.146728 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9\" (UID: \"56b1c51b-bdcd-4c55-a00d-105ab4bc04b0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9" Oct 01 10:16:08 crc kubenswrapper[4787]: I1001 10:16:08.147146 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9\" (UID: \"56b1c51b-bdcd-4c55-a00d-105ab4bc04b0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9" Oct 01 10:16:08 crc kubenswrapper[4787]: I1001 10:16:08.147194 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9\" (UID: \"56b1c51b-bdcd-4c55-a00d-105ab4bc04b0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9" Oct 01 10:16:08 crc kubenswrapper[4787]: I1001 10:16:08.147271 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9\" (UID: \"56b1c51b-bdcd-4c55-a00d-105ab4bc04b0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9" Oct 01 10:16:08 crc kubenswrapper[4787]: I1001 10:16:08.165189 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42g4q\" (UniqueName: \"kubernetes.io/projected/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-kube-api-access-42g4q\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9\" (UID: \"56b1c51b-bdcd-4c55-a00d-105ab4bc04b0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9" Oct 01 10:16:08 crc kubenswrapper[4787]: I1001 10:16:08.224888 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9" Oct 01 10:16:08 crc kubenswrapper[4787]: I1001 10:16:08.796586 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9"] Oct 01 10:16:09 crc kubenswrapper[4787]: I1001 10:16:09.607742 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 01 10:16:09 crc kubenswrapper[4787]: I1001 10:16:09.815539 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9" event={"ID":"56b1c51b-bdcd-4c55-a00d-105ab4bc04b0","Type":"ContainerStarted","Data":"d7085d6f7fd932fe8cac79e2d664e2549e39e30ff2945a3fd472264b4bc9b0b5"} Oct 01 10:16:10 crc kubenswrapper[4787]: I1001 10:16:10.826351 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9" event={"ID":"56b1c51b-bdcd-4c55-a00d-105ab4bc04b0","Type":"ContainerStarted","Data":"58d29bf982548840111c729d82fdacf1174eed3f3fcfd7791b4e00219cb15878"} Oct 01 10:16:10 crc kubenswrapper[4787]: I1001 10:16:10.856166 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9" podStartSLOduration=3.053472309 podStartE2EDuration="3.856138189s" podCreationTimestamp="2025-10-01 10:16:07 +0000 UTC" firstStartedPulling="2025-10-01 10:16:08.802202362 +0000 UTC m=+2400.917346519" lastFinishedPulling="2025-10-01 10:16:09.604868222 +0000 UTC m=+2401.720012399" observedRunningTime="2025-10-01 10:16:10.84174859 +0000 UTC m=+2402.956892767" watchObservedRunningTime="2025-10-01 10:16:10.856138189 +0000 UTC m=+2402.971282386" Oct 01 10:16:19 crc kubenswrapper[4787]: I1001 10:16:19.524574 4787 scope.go:117] "RemoveContainer" containerID="e05f5ad393ada7db54df5c58983c6c4e3e27cad1d8e65e9204e8b48c5e3ba7c1" Oct 01 10:16:19 crc kubenswrapper[4787]: E1001 10:16:19.525614 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:16:31 crc kubenswrapper[4787]: I1001 10:16:31.524218 4787 scope.go:117] "RemoveContainer" containerID="e05f5ad393ada7db54df5c58983c6c4e3e27cad1d8e65e9204e8b48c5e3ba7c1" Oct 01 10:16:31 crc kubenswrapper[4787]: E1001 10:16:31.525187 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:16:43 crc kubenswrapper[4787]: I1001 10:16:43.523554 4787 scope.go:117] "RemoveContainer" containerID="e05f5ad393ada7db54df5c58983c6c4e3e27cad1d8e65e9204e8b48c5e3ba7c1" Oct 01 10:16:43 crc kubenswrapper[4787]: E1001 10:16:43.524239 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:16:55 crc kubenswrapper[4787]: I1001 10:16:55.524210 4787 scope.go:117] "RemoveContainer" containerID="e05f5ad393ada7db54df5c58983c6c4e3e27cad1d8e65e9204e8b48c5e3ba7c1" Oct 01 10:16:55 crc kubenswrapper[4787]: E1001 10:16:55.525975 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:17:10 crc kubenswrapper[4787]: I1001 10:17:10.524107 4787 scope.go:117] "RemoveContainer" containerID="e05f5ad393ada7db54df5c58983c6c4e3e27cad1d8e65e9204e8b48c5e3ba7c1" Oct 01 10:17:10 crc kubenswrapper[4787]: E1001 10:17:10.524862 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:17:21 crc kubenswrapper[4787]: I1001 10:17:21.524632 4787 scope.go:117] "RemoveContainer" containerID="e05f5ad393ada7db54df5c58983c6c4e3e27cad1d8e65e9204e8b48c5e3ba7c1" Oct 01 10:17:22 crc kubenswrapper[4787]: I1001 10:17:22.535194 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" event={"ID":"fa51b883-5f3e-4141-9d4c-37704aac7718","Type":"ContainerStarted","Data":"f6c4f2a4060e1ca1135f9288e823ff4cd4277a621cbd921d70bcb532fa4e420a"} Oct 01 10:18:44 crc kubenswrapper[4787]: I1001 10:18:44.301839 4787 generic.go:334] "Generic (PLEG): container finished" podID="56b1c51b-bdcd-4c55-a00d-105ab4bc04b0" containerID="58d29bf982548840111c729d82fdacf1174eed3f3fcfd7791b4e00219cb15878" exitCode=0 Oct 01 10:18:44 crc kubenswrapper[4787]: I1001 10:18:44.301911 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9" event={"ID":"56b1c51b-bdcd-4c55-a00d-105ab4bc04b0","Type":"ContainerDied","Data":"58d29bf982548840111c729d82fdacf1174eed3f3fcfd7791b4e00219cb15878"} Oct 01 10:18:45 crc kubenswrapper[4787]: I1001 10:18:45.751230 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9" Oct 01 10:18:45 crc kubenswrapper[4787]: I1001 10:18:45.946165 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-ceilometer-compute-config-data-2\") pod \"56b1c51b-bdcd-4c55-a00d-105ab4bc04b0\" (UID: \"56b1c51b-bdcd-4c55-a00d-105ab4bc04b0\") " Oct 01 10:18:45 crc kubenswrapper[4787]: I1001 10:18:45.946483 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-telemetry-combined-ca-bundle\") pod \"56b1c51b-bdcd-4c55-a00d-105ab4bc04b0\" (UID: \"56b1c51b-bdcd-4c55-a00d-105ab4bc04b0\") " Oct 01 10:18:45 crc kubenswrapper[4787]: I1001 10:18:45.946544 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-ceilometer-compute-config-data-0\") pod \"56b1c51b-bdcd-4c55-a00d-105ab4bc04b0\" (UID: \"56b1c51b-bdcd-4c55-a00d-105ab4bc04b0\") " Oct 01 10:18:45 crc kubenswrapper[4787]: I1001 10:18:45.946620 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-ssh-key\") pod \"56b1c51b-bdcd-4c55-a00d-105ab4bc04b0\" (UID: \"56b1c51b-bdcd-4c55-a00d-105ab4bc04b0\") " Oct 01 10:18:45 crc kubenswrapper[4787]: I1001 10:18:45.946692 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-inventory\") pod \"56b1c51b-bdcd-4c55-a00d-105ab4bc04b0\" (UID: \"56b1c51b-bdcd-4c55-a00d-105ab4bc04b0\") " Oct 01 10:18:45 crc kubenswrapper[4787]: I1001 10:18:45.946735 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42g4q\" (UniqueName: \"kubernetes.io/projected/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-kube-api-access-42g4q\") pod \"56b1c51b-bdcd-4c55-a00d-105ab4bc04b0\" (UID: \"56b1c51b-bdcd-4c55-a00d-105ab4bc04b0\") " Oct 01 10:18:45 crc kubenswrapper[4787]: I1001 10:18:45.946804 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-ceilometer-compute-config-data-1\") pod \"56b1c51b-bdcd-4c55-a00d-105ab4bc04b0\" (UID: \"56b1c51b-bdcd-4c55-a00d-105ab4bc04b0\") " Oct 01 10:18:45 crc kubenswrapper[4787]: I1001 10:18:45.954132 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-kube-api-access-42g4q" (OuterVolumeSpecName: "kube-api-access-42g4q") pod "56b1c51b-bdcd-4c55-a00d-105ab4bc04b0" (UID: "56b1c51b-bdcd-4c55-a00d-105ab4bc04b0"). InnerVolumeSpecName "kube-api-access-42g4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:18:45 crc kubenswrapper[4787]: I1001 10:18:45.960315 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "56b1c51b-bdcd-4c55-a00d-105ab4bc04b0" (UID: "56b1c51b-bdcd-4c55-a00d-105ab4bc04b0"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:18:45 crc kubenswrapper[4787]: I1001 10:18:45.979192 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-inventory" (OuterVolumeSpecName: "inventory") pod "56b1c51b-bdcd-4c55-a00d-105ab4bc04b0" (UID: "56b1c51b-bdcd-4c55-a00d-105ab4bc04b0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:18:45 crc kubenswrapper[4787]: I1001 10:18:45.979536 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "56b1c51b-bdcd-4c55-a00d-105ab4bc04b0" (UID: "56b1c51b-bdcd-4c55-a00d-105ab4bc04b0"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:18:45 crc kubenswrapper[4787]: I1001 10:18:45.983117 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "56b1c51b-bdcd-4c55-a00d-105ab4bc04b0" (UID: "56b1c51b-bdcd-4c55-a00d-105ab4bc04b0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:18:45 crc kubenswrapper[4787]: I1001 10:18:45.992982 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "56b1c51b-bdcd-4c55-a00d-105ab4bc04b0" (UID: "56b1c51b-bdcd-4c55-a00d-105ab4bc04b0"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:18:46 crc kubenswrapper[4787]: I1001 10:18:46.000316 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "56b1c51b-bdcd-4c55-a00d-105ab4bc04b0" (UID: "56b1c51b-bdcd-4c55-a00d-105ab4bc04b0"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:18:46 crc kubenswrapper[4787]: I1001 10:18:46.049275 4787 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Oct 01 10:18:46 crc kubenswrapper[4787]: I1001 10:18:46.049327 4787 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Oct 01 10:18:46 crc kubenswrapper[4787]: I1001 10:18:46.049348 4787 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 01 10:18:46 crc kubenswrapper[4787]: I1001 10:18:46.049367 4787 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Oct 01 10:18:46 crc kubenswrapper[4787]: I1001 10:18:46.049389 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 10:18:46 crc kubenswrapper[4787]: I1001 10:18:46.049408 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-inventory\") on node \"crc\" DevicePath \"\"" Oct 01 10:18:46 crc kubenswrapper[4787]: I1001 10:18:46.049449 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42g4q\" (UniqueName: \"kubernetes.io/projected/56b1c51b-bdcd-4c55-a00d-105ab4bc04b0-kube-api-access-42g4q\") on node \"crc\" DevicePath \"\"" Oct 01 10:18:46 crc kubenswrapper[4787]: I1001 10:18:46.378633 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9" event={"ID":"56b1c51b-bdcd-4c55-a00d-105ab4bc04b0","Type":"ContainerDied","Data":"d7085d6f7fd932fe8cac79e2d664e2549e39e30ff2945a3fd472264b4bc9b0b5"} Oct 01 10:18:46 crc kubenswrapper[4787]: I1001 10:18:46.378688 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7085d6f7fd932fe8cac79e2d664e2549e39e30ff2945a3fd472264b4bc9b0b5" Oct 01 10:18:46 crc kubenswrapper[4787]: I1001 10:18:46.378781 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9" Oct 01 10:19:13 crc kubenswrapper[4787]: I1001 10:19:13.376643 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-5c7b59dd4f-spxm4" podUID="e743cd5e-e8c6-4fe2-9480-3a30316b8e23" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Oct 01 10:19:29 crc kubenswrapper[4787]: I1001 10:19:29.146402 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Oct 01 10:19:29 crc kubenswrapper[4787]: E1001 10:19:29.147816 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56b1c51b-bdcd-4c55-a00d-105ab4bc04b0" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 01 10:19:29 crc kubenswrapper[4787]: I1001 10:19:29.147842 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="56b1c51b-bdcd-4c55-a00d-105ab4bc04b0" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 01 10:19:29 crc kubenswrapper[4787]: I1001 10:19:29.148279 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="56b1c51b-bdcd-4c55-a00d-105ab4bc04b0" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 01 10:19:29 crc kubenswrapper[4787]: I1001 10:19:29.149514 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 01 10:19:29 crc kubenswrapper[4787]: I1001 10:19:29.152114 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Oct 01 10:19:29 crc kubenswrapper[4787]: I1001 10:19:29.152175 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-qq94k" Oct 01 10:19:29 crc kubenswrapper[4787]: I1001 10:19:29.152486 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Oct 01 10:19:29 crc kubenswrapper[4787]: I1001 10:19:29.153532 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 01 10:19:29 crc kubenswrapper[4787]: I1001 10:19:29.155985 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 01 10:19:29 crc kubenswrapper[4787]: I1001 10:19:29.289478 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/4b136104-8797-40d5-9ea1-fec8c6be0e90-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"4b136104-8797-40d5-9ea1-fec8c6be0e90\") " pod="openstack/tempest-tests-tempest" Oct 01 10:19:29 crc kubenswrapper[4787]: I1001 10:19:29.289700 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4b136104-8797-40d5-9ea1-fec8c6be0e90-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"4b136104-8797-40d5-9ea1-fec8c6be0e90\") " pod="openstack/tempest-tests-tempest" Oct 01 10:19:29 crc kubenswrapper[4787]: I1001 10:19:29.289757 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4b136104-8797-40d5-9ea1-fec8c6be0e90-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"4b136104-8797-40d5-9ea1-fec8c6be0e90\") " pod="openstack/tempest-tests-tempest" Oct 01 10:19:29 crc kubenswrapper[4787]: I1001 10:19:29.289876 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gllb6\" (UniqueName: \"kubernetes.io/projected/4b136104-8797-40d5-9ea1-fec8c6be0e90-kube-api-access-gllb6\") pod \"tempest-tests-tempest\" (UID: \"4b136104-8797-40d5-9ea1-fec8c6be0e90\") " pod="openstack/tempest-tests-tempest" Oct 01 10:19:29 crc kubenswrapper[4787]: I1001 10:19:29.289978 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/4b136104-8797-40d5-9ea1-fec8c6be0e90-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"4b136104-8797-40d5-9ea1-fec8c6be0e90\") " pod="openstack/tempest-tests-tempest" Oct 01 10:19:29 crc kubenswrapper[4787]: I1001 10:19:29.290133 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b136104-8797-40d5-9ea1-fec8c6be0e90-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"4b136104-8797-40d5-9ea1-fec8c6be0e90\") " pod="openstack/tempest-tests-tempest" Oct 01 10:19:29 crc kubenswrapper[4787]: I1001 10:19:29.290195 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4b136104-8797-40d5-9ea1-fec8c6be0e90-config-data\") pod \"tempest-tests-tempest\" (UID: \"4b136104-8797-40d5-9ea1-fec8c6be0e90\") " pod="openstack/tempest-tests-tempest" Oct 01 10:19:29 crc kubenswrapper[4787]: I1001 10:19:29.290265 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest\" (UID: \"4b136104-8797-40d5-9ea1-fec8c6be0e90\") " pod="openstack/tempest-tests-tempest" Oct 01 10:19:29 crc kubenswrapper[4787]: I1001 10:19:29.290489 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/4b136104-8797-40d5-9ea1-fec8c6be0e90-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"4b136104-8797-40d5-9ea1-fec8c6be0e90\") " pod="openstack/tempest-tests-tempest" Oct 01 10:19:29 crc kubenswrapper[4787]: I1001 10:19:29.392972 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/4b136104-8797-40d5-9ea1-fec8c6be0e90-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"4b136104-8797-40d5-9ea1-fec8c6be0e90\") " pod="openstack/tempest-tests-tempest" Oct 01 10:19:29 crc kubenswrapper[4787]: I1001 10:19:29.393846 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4b136104-8797-40d5-9ea1-fec8c6be0e90-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"4b136104-8797-40d5-9ea1-fec8c6be0e90\") " pod="openstack/tempest-tests-tempest" Oct 01 10:19:29 crc kubenswrapper[4787]: I1001 10:19:29.393498 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/4b136104-8797-40d5-9ea1-fec8c6be0e90-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"4b136104-8797-40d5-9ea1-fec8c6be0e90\") " pod="openstack/tempest-tests-tempest" Oct 01 10:19:29 crc kubenswrapper[4787]: I1001 10:19:29.393897 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4b136104-8797-40d5-9ea1-fec8c6be0e90-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"4b136104-8797-40d5-9ea1-fec8c6be0e90\") " pod="openstack/tempest-tests-tempest" Oct 01 10:19:29 crc kubenswrapper[4787]: I1001 10:19:29.393995 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gllb6\" (UniqueName: \"kubernetes.io/projected/4b136104-8797-40d5-9ea1-fec8c6be0e90-kube-api-access-gllb6\") pod \"tempest-tests-tempest\" (UID: \"4b136104-8797-40d5-9ea1-fec8c6be0e90\") " pod="openstack/tempest-tests-tempest" Oct 01 10:19:29 crc kubenswrapper[4787]: I1001 10:19:29.394101 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/4b136104-8797-40d5-9ea1-fec8c6be0e90-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"4b136104-8797-40d5-9ea1-fec8c6be0e90\") " pod="openstack/tempest-tests-tempest" Oct 01 10:19:29 crc kubenswrapper[4787]: I1001 10:19:29.394203 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b136104-8797-40d5-9ea1-fec8c6be0e90-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"4b136104-8797-40d5-9ea1-fec8c6be0e90\") " pod="openstack/tempest-tests-tempest" Oct 01 10:19:29 crc kubenswrapper[4787]: I1001 10:19:29.394238 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4b136104-8797-40d5-9ea1-fec8c6be0e90-config-data\") pod \"tempest-tests-tempest\" (UID: \"4b136104-8797-40d5-9ea1-fec8c6be0e90\") " pod="openstack/tempest-tests-tempest" Oct 01 10:19:29 crc kubenswrapper[4787]: I1001 10:19:29.394300 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest\" (UID: \"4b136104-8797-40d5-9ea1-fec8c6be0e90\") " pod="openstack/tempest-tests-tempest" Oct 01 10:19:29 crc kubenswrapper[4787]: I1001 10:19:29.394549 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/4b136104-8797-40d5-9ea1-fec8c6be0e90-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"4b136104-8797-40d5-9ea1-fec8c6be0e90\") " pod="openstack/tempest-tests-tempest" Oct 01 10:19:29 crc kubenswrapper[4787]: I1001 10:19:29.394696 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4b136104-8797-40d5-9ea1-fec8c6be0e90-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"4b136104-8797-40d5-9ea1-fec8c6be0e90\") " pod="openstack/tempest-tests-tempest" Oct 01 10:19:29 crc kubenswrapper[4787]: I1001 10:19:29.395406 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest\" (UID: \"4b136104-8797-40d5-9ea1-fec8c6be0e90\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/tempest-tests-tempest" Oct 01 10:19:29 crc kubenswrapper[4787]: I1001 10:19:29.395435 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/4b136104-8797-40d5-9ea1-fec8c6be0e90-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"4b136104-8797-40d5-9ea1-fec8c6be0e90\") " pod="openstack/tempest-tests-tempest" Oct 01 10:19:29 crc kubenswrapper[4787]: I1001 10:19:29.396249 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4b136104-8797-40d5-9ea1-fec8c6be0e90-config-data\") pod \"tempest-tests-tempest\" (UID: \"4b136104-8797-40d5-9ea1-fec8c6be0e90\") " pod="openstack/tempest-tests-tempest" Oct 01 10:19:29 crc kubenswrapper[4787]: I1001 10:19:29.402351 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b136104-8797-40d5-9ea1-fec8c6be0e90-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"4b136104-8797-40d5-9ea1-fec8c6be0e90\") " pod="openstack/tempest-tests-tempest" Oct 01 10:19:29 crc kubenswrapper[4787]: I1001 10:19:29.402374 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4b136104-8797-40d5-9ea1-fec8c6be0e90-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"4b136104-8797-40d5-9ea1-fec8c6be0e90\") " pod="openstack/tempest-tests-tempest" Oct 01 10:19:29 crc kubenswrapper[4787]: I1001 10:19:29.413744 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/4b136104-8797-40d5-9ea1-fec8c6be0e90-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"4b136104-8797-40d5-9ea1-fec8c6be0e90\") " pod="openstack/tempest-tests-tempest" Oct 01 10:19:29 crc kubenswrapper[4787]: I1001 10:19:29.418468 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gllb6\" (UniqueName: \"kubernetes.io/projected/4b136104-8797-40d5-9ea1-fec8c6be0e90-kube-api-access-gllb6\") pod \"tempest-tests-tempest\" (UID: \"4b136104-8797-40d5-9ea1-fec8c6be0e90\") " pod="openstack/tempest-tests-tempest" Oct 01 10:19:29 crc kubenswrapper[4787]: I1001 10:19:29.450749 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest\" (UID: \"4b136104-8797-40d5-9ea1-fec8c6be0e90\") " pod="openstack/tempest-tests-tempest" Oct 01 10:19:29 crc kubenswrapper[4787]: I1001 10:19:29.507510 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 01 10:19:29 crc kubenswrapper[4787]: I1001 10:19:29.989760 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 01 10:19:29 crc kubenswrapper[4787]: W1001 10:19:29.992136 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b136104_8797_40d5_9ea1_fec8c6be0e90.slice/crio-6042ac4649ddff4357aa290f8de50125f01a705ec06864e6086fe9777317fa00 WatchSource:0}: Error finding container 6042ac4649ddff4357aa290f8de50125f01a705ec06864e6086fe9777317fa00: Status 404 returned error can't find the container with id 6042ac4649ddff4357aa290f8de50125f01a705ec06864e6086fe9777317fa00 Oct 01 10:19:29 crc kubenswrapper[4787]: I1001 10:19:29.996875 4787 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 10:19:30 crc kubenswrapper[4787]: I1001 10:19:30.832521 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"4b136104-8797-40d5-9ea1-fec8c6be0e90","Type":"ContainerStarted","Data":"6042ac4649ddff4357aa290f8de50125f01a705ec06864e6086fe9777317fa00"} Oct 01 10:19:41 crc kubenswrapper[4787]: I1001 10:19:41.250964 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:19:41 crc kubenswrapper[4787]: I1001 10:19:41.251637 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:19:57 crc kubenswrapper[4787]: E1001 10:19:57.592062 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Oct 01 10:19:57 crc kubenswrapper[4787]: E1001 10:19:57.592870 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gllb6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(4b136104-8797-40d5-9ea1-fec8c6be0e90): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 01 10:19:57 crc kubenswrapper[4787]: E1001 10:19:57.594146 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="4b136104-8797-40d5-9ea1-fec8c6be0e90" Oct 01 10:19:58 crc kubenswrapper[4787]: E1001 10:19:58.147446 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="4b136104-8797-40d5-9ea1-fec8c6be0e90" Oct 01 10:20:10 crc kubenswrapper[4787]: I1001 10:20:10.999111 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 01 10:20:11 crc kubenswrapper[4787]: I1001 10:20:11.251284 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:20:11 crc kubenswrapper[4787]: I1001 10:20:11.251749 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:20:12 crc kubenswrapper[4787]: I1001 10:20:12.285524 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"4b136104-8797-40d5-9ea1-fec8c6be0e90","Type":"ContainerStarted","Data":"1221312c3fea241a9c775a3b4f5142913e44ed581fc20935359dbbd60a68af15"} Oct 01 10:20:12 crc kubenswrapper[4787]: I1001 10:20:12.301841 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.302202605 podStartE2EDuration="44.301810994s" podCreationTimestamp="2025-10-01 10:19:28 +0000 UTC" firstStartedPulling="2025-10-01 10:19:29.996576418 +0000 UTC m=+2602.111720575" lastFinishedPulling="2025-10-01 10:20:10.996184797 +0000 UTC m=+2643.111328964" observedRunningTime="2025-10-01 10:20:12.300276147 +0000 UTC m=+2644.415420314" watchObservedRunningTime="2025-10-01 10:20:12.301810994 +0000 UTC m=+2644.416955171" Oct 01 10:20:41 crc kubenswrapper[4787]: I1001 10:20:41.250432 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:20:41 crc kubenswrapper[4787]: I1001 10:20:41.250973 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:20:41 crc kubenswrapper[4787]: I1001 10:20:41.251013 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" Oct 01 10:20:41 crc kubenswrapper[4787]: I1001 10:20:41.251576 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f6c4f2a4060e1ca1135f9288e823ff4cd4277a621cbd921d70bcb532fa4e420a"} pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 10:20:41 crc kubenswrapper[4787]: I1001 10:20:41.251662 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" containerID="cri-o://f6c4f2a4060e1ca1135f9288e823ff4cd4277a621cbd921d70bcb532fa4e420a" gracePeriod=600 Oct 01 10:20:41 crc kubenswrapper[4787]: I1001 10:20:41.568590 4787 generic.go:334] "Generic (PLEG): container finished" podID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerID="f6c4f2a4060e1ca1135f9288e823ff4cd4277a621cbd921d70bcb532fa4e420a" exitCode=0 Oct 01 10:20:41 crc kubenswrapper[4787]: I1001 10:20:41.568662 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" event={"ID":"fa51b883-5f3e-4141-9d4c-37704aac7718","Type":"ContainerDied","Data":"f6c4f2a4060e1ca1135f9288e823ff4cd4277a621cbd921d70bcb532fa4e420a"} Oct 01 10:20:41 crc kubenswrapper[4787]: I1001 10:20:41.568896 4787 scope.go:117] "RemoveContainer" containerID="e05f5ad393ada7db54df5c58983c6c4e3e27cad1d8e65e9204e8b48c5e3ba7c1" Oct 01 10:20:42 crc kubenswrapper[4787]: I1001 10:20:42.597046 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" event={"ID":"fa51b883-5f3e-4141-9d4c-37704aac7718","Type":"ContainerStarted","Data":"3df952eaa93a8322715e6a8d23a95eedf91662575215810014140a3e7d2a03bd"} Oct 01 10:20:43 crc kubenswrapper[4787]: I1001 10:20:43.220628 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-blbn4"] Oct 01 10:20:43 crc kubenswrapper[4787]: I1001 10:20:43.223151 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-blbn4" Oct 01 10:20:43 crc kubenswrapper[4787]: I1001 10:20:43.254155 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-blbn4"] Oct 01 10:20:43 crc kubenswrapper[4787]: I1001 10:20:43.296786 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbd816a4-c3b2-4a94-9e2a-986a24503460-utilities\") pod \"community-operators-blbn4\" (UID: \"dbd816a4-c3b2-4a94-9e2a-986a24503460\") " pod="openshift-marketplace/community-operators-blbn4" Oct 01 10:20:43 crc kubenswrapper[4787]: I1001 10:20:43.296883 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbd816a4-c3b2-4a94-9e2a-986a24503460-catalog-content\") pod \"community-operators-blbn4\" (UID: \"dbd816a4-c3b2-4a94-9e2a-986a24503460\") " pod="openshift-marketplace/community-operators-blbn4" Oct 01 10:20:43 crc kubenswrapper[4787]: I1001 10:20:43.296914 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgr89\" (UniqueName: \"kubernetes.io/projected/dbd816a4-c3b2-4a94-9e2a-986a24503460-kube-api-access-hgr89\") pod \"community-operators-blbn4\" (UID: \"dbd816a4-c3b2-4a94-9e2a-986a24503460\") " pod="openshift-marketplace/community-operators-blbn4" Oct 01 10:20:43 crc kubenswrapper[4787]: I1001 10:20:43.398765 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbd816a4-c3b2-4a94-9e2a-986a24503460-utilities\") pod \"community-operators-blbn4\" (UID: \"dbd816a4-c3b2-4a94-9e2a-986a24503460\") " pod="openshift-marketplace/community-operators-blbn4" Oct 01 10:20:43 crc kubenswrapper[4787]: I1001 10:20:43.399144 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbd816a4-c3b2-4a94-9e2a-986a24503460-catalog-content\") pod \"community-operators-blbn4\" (UID: \"dbd816a4-c3b2-4a94-9e2a-986a24503460\") " pod="openshift-marketplace/community-operators-blbn4" Oct 01 10:20:43 crc kubenswrapper[4787]: I1001 10:20:43.399174 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgr89\" (UniqueName: \"kubernetes.io/projected/dbd816a4-c3b2-4a94-9e2a-986a24503460-kube-api-access-hgr89\") pod \"community-operators-blbn4\" (UID: \"dbd816a4-c3b2-4a94-9e2a-986a24503460\") " pod="openshift-marketplace/community-operators-blbn4" Oct 01 10:20:43 crc kubenswrapper[4787]: I1001 10:20:43.399610 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbd816a4-c3b2-4a94-9e2a-986a24503460-utilities\") pod \"community-operators-blbn4\" (UID: \"dbd816a4-c3b2-4a94-9e2a-986a24503460\") " pod="openshift-marketplace/community-operators-blbn4" Oct 01 10:20:43 crc kubenswrapper[4787]: I1001 10:20:43.399747 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbd816a4-c3b2-4a94-9e2a-986a24503460-catalog-content\") pod \"community-operators-blbn4\" (UID: \"dbd816a4-c3b2-4a94-9e2a-986a24503460\") " pod="openshift-marketplace/community-operators-blbn4" Oct 01 10:20:43 crc kubenswrapper[4787]: I1001 10:20:43.419207 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgr89\" (UniqueName: \"kubernetes.io/projected/dbd816a4-c3b2-4a94-9e2a-986a24503460-kube-api-access-hgr89\") pod \"community-operators-blbn4\" (UID: \"dbd816a4-c3b2-4a94-9e2a-986a24503460\") " pod="openshift-marketplace/community-operators-blbn4" Oct 01 10:20:43 crc kubenswrapper[4787]: I1001 10:20:43.553065 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-blbn4" Oct 01 10:20:44 crc kubenswrapper[4787]: I1001 10:20:44.036833 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-blbn4"] Oct 01 10:20:44 crc kubenswrapper[4787]: I1001 10:20:44.627860 4787 generic.go:334] "Generic (PLEG): container finished" podID="dbd816a4-c3b2-4a94-9e2a-986a24503460" containerID="0dcb68f3cd52adb3a8624753edecac15d5b29859d4ab0b4b3ea07208f4d884f8" exitCode=0 Oct 01 10:20:44 crc kubenswrapper[4787]: I1001 10:20:44.628244 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-blbn4" event={"ID":"dbd816a4-c3b2-4a94-9e2a-986a24503460","Type":"ContainerDied","Data":"0dcb68f3cd52adb3a8624753edecac15d5b29859d4ab0b4b3ea07208f4d884f8"} Oct 01 10:20:44 crc kubenswrapper[4787]: I1001 10:20:44.628280 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-blbn4" event={"ID":"dbd816a4-c3b2-4a94-9e2a-986a24503460","Type":"ContainerStarted","Data":"275ca60dd3374b8357ef7a3d7e57a9297e82d1bfbdf0886cc3e6a2266291044e"} Oct 01 10:20:45 crc kubenswrapper[4787]: I1001 10:20:45.431953 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ptk7v"] Oct 01 10:20:45 crc kubenswrapper[4787]: I1001 10:20:45.434579 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ptk7v" Oct 01 10:20:45 crc kubenswrapper[4787]: I1001 10:20:45.444331 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ptk7v"] Oct 01 10:20:45 crc kubenswrapper[4787]: I1001 10:20:45.567016 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ccf3825-2cc8-491a-9a37-1973aa0dba30-utilities\") pod \"redhat-operators-ptk7v\" (UID: \"9ccf3825-2cc8-491a-9a37-1973aa0dba30\") " pod="openshift-marketplace/redhat-operators-ptk7v" Oct 01 10:20:45 crc kubenswrapper[4787]: I1001 10:20:45.567390 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ccf3825-2cc8-491a-9a37-1973aa0dba30-catalog-content\") pod \"redhat-operators-ptk7v\" (UID: \"9ccf3825-2cc8-491a-9a37-1973aa0dba30\") " pod="openshift-marketplace/redhat-operators-ptk7v" Oct 01 10:20:45 crc kubenswrapper[4787]: I1001 10:20:45.567563 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffptl\" (UniqueName: \"kubernetes.io/projected/9ccf3825-2cc8-491a-9a37-1973aa0dba30-kube-api-access-ffptl\") pod \"redhat-operators-ptk7v\" (UID: \"9ccf3825-2cc8-491a-9a37-1973aa0dba30\") " pod="openshift-marketplace/redhat-operators-ptk7v" Oct 01 10:20:45 crc kubenswrapper[4787]: I1001 10:20:45.617851 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-h4952"] Oct 01 10:20:45 crc kubenswrapper[4787]: I1001 10:20:45.620436 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h4952" Oct 01 10:20:45 crc kubenswrapper[4787]: I1001 10:20:45.642235 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h4952"] Oct 01 10:20:45 crc kubenswrapper[4787]: I1001 10:20:45.670993 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce0e38fc-4300-4dd8-add8-4316579c9617-catalog-content\") pod \"certified-operators-h4952\" (UID: \"ce0e38fc-4300-4dd8-add8-4316579c9617\") " pod="openshift-marketplace/certified-operators-h4952" Oct 01 10:20:45 crc kubenswrapper[4787]: I1001 10:20:45.671115 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ccf3825-2cc8-491a-9a37-1973aa0dba30-utilities\") pod \"redhat-operators-ptk7v\" (UID: \"9ccf3825-2cc8-491a-9a37-1973aa0dba30\") " pod="openshift-marketplace/redhat-operators-ptk7v" Oct 01 10:20:45 crc kubenswrapper[4787]: I1001 10:20:45.671176 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ccf3825-2cc8-491a-9a37-1973aa0dba30-catalog-content\") pod \"redhat-operators-ptk7v\" (UID: \"9ccf3825-2cc8-491a-9a37-1973aa0dba30\") " pod="openshift-marketplace/redhat-operators-ptk7v" Oct 01 10:20:45 crc kubenswrapper[4787]: I1001 10:20:45.671208 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsjjs\" (UniqueName: \"kubernetes.io/projected/ce0e38fc-4300-4dd8-add8-4316579c9617-kube-api-access-vsjjs\") pod \"certified-operators-h4952\" (UID: \"ce0e38fc-4300-4dd8-add8-4316579c9617\") " pod="openshift-marketplace/certified-operators-h4952" Oct 01 10:20:45 crc kubenswrapper[4787]: I1001 10:20:45.671256 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce0e38fc-4300-4dd8-add8-4316579c9617-utilities\") pod \"certified-operators-h4952\" (UID: \"ce0e38fc-4300-4dd8-add8-4316579c9617\") " pod="openshift-marketplace/certified-operators-h4952" Oct 01 10:20:45 crc kubenswrapper[4787]: I1001 10:20:45.671273 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffptl\" (UniqueName: \"kubernetes.io/projected/9ccf3825-2cc8-491a-9a37-1973aa0dba30-kube-api-access-ffptl\") pod \"redhat-operators-ptk7v\" (UID: \"9ccf3825-2cc8-491a-9a37-1973aa0dba30\") " pod="openshift-marketplace/redhat-operators-ptk7v" Oct 01 10:20:45 crc kubenswrapper[4787]: I1001 10:20:45.671926 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ccf3825-2cc8-491a-9a37-1973aa0dba30-utilities\") pod \"redhat-operators-ptk7v\" (UID: \"9ccf3825-2cc8-491a-9a37-1973aa0dba30\") " pod="openshift-marketplace/redhat-operators-ptk7v" Oct 01 10:20:45 crc kubenswrapper[4787]: I1001 10:20:45.672344 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ccf3825-2cc8-491a-9a37-1973aa0dba30-catalog-content\") pod \"redhat-operators-ptk7v\" (UID: \"9ccf3825-2cc8-491a-9a37-1973aa0dba30\") " pod="openshift-marketplace/redhat-operators-ptk7v" Oct 01 10:20:45 crc kubenswrapper[4787]: I1001 10:20:45.697877 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffptl\" (UniqueName: \"kubernetes.io/projected/9ccf3825-2cc8-491a-9a37-1973aa0dba30-kube-api-access-ffptl\") pod \"redhat-operators-ptk7v\" (UID: \"9ccf3825-2cc8-491a-9a37-1973aa0dba30\") " pod="openshift-marketplace/redhat-operators-ptk7v" Oct 01 10:20:45 crc kubenswrapper[4787]: I1001 10:20:45.757030 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ptk7v" Oct 01 10:20:45 crc kubenswrapper[4787]: I1001 10:20:45.777281 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce0e38fc-4300-4dd8-add8-4316579c9617-catalog-content\") pod \"certified-operators-h4952\" (UID: \"ce0e38fc-4300-4dd8-add8-4316579c9617\") " pod="openshift-marketplace/certified-operators-h4952" Oct 01 10:20:45 crc kubenswrapper[4787]: I1001 10:20:45.777395 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsjjs\" (UniqueName: \"kubernetes.io/projected/ce0e38fc-4300-4dd8-add8-4316579c9617-kube-api-access-vsjjs\") pod \"certified-operators-h4952\" (UID: \"ce0e38fc-4300-4dd8-add8-4316579c9617\") " pod="openshift-marketplace/certified-operators-h4952" Oct 01 10:20:45 crc kubenswrapper[4787]: I1001 10:20:45.777440 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce0e38fc-4300-4dd8-add8-4316579c9617-utilities\") pod \"certified-operators-h4952\" (UID: \"ce0e38fc-4300-4dd8-add8-4316579c9617\") " pod="openshift-marketplace/certified-operators-h4952" Oct 01 10:20:45 crc kubenswrapper[4787]: I1001 10:20:45.778375 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce0e38fc-4300-4dd8-add8-4316579c9617-catalog-content\") pod \"certified-operators-h4952\" (UID: \"ce0e38fc-4300-4dd8-add8-4316579c9617\") " pod="openshift-marketplace/certified-operators-h4952" Oct 01 10:20:45 crc kubenswrapper[4787]: I1001 10:20:45.778462 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce0e38fc-4300-4dd8-add8-4316579c9617-utilities\") pod \"certified-operators-h4952\" (UID: \"ce0e38fc-4300-4dd8-add8-4316579c9617\") " pod="openshift-marketplace/certified-operators-h4952" Oct 01 10:20:45 crc kubenswrapper[4787]: I1001 10:20:45.801425 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsjjs\" (UniqueName: \"kubernetes.io/projected/ce0e38fc-4300-4dd8-add8-4316579c9617-kube-api-access-vsjjs\") pod \"certified-operators-h4952\" (UID: \"ce0e38fc-4300-4dd8-add8-4316579c9617\") " pod="openshift-marketplace/certified-operators-h4952" Oct 01 10:20:45 crc kubenswrapper[4787]: I1001 10:20:45.942654 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h4952" Oct 01 10:20:46 crc kubenswrapper[4787]: I1001 10:20:46.288616 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ptk7v"] Oct 01 10:20:46 crc kubenswrapper[4787]: I1001 10:20:46.579212 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h4952"] Oct 01 10:20:46 crc kubenswrapper[4787]: I1001 10:20:46.665912 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h4952" event={"ID":"ce0e38fc-4300-4dd8-add8-4316579c9617","Type":"ContainerStarted","Data":"3f41a7b5f466f643ee3ea454959dec54cb68be7dc5312a32803876d5c24f09b9"} Oct 01 10:20:46 crc kubenswrapper[4787]: I1001 10:20:46.671726 4787 generic.go:334] "Generic (PLEG): container finished" podID="9ccf3825-2cc8-491a-9a37-1973aa0dba30" containerID="1ddb5f4d9acabf0d18435e45a5c896643f11d57078145a7b7edbd9fdf8048087" exitCode=0 Oct 01 10:20:46 crc kubenswrapper[4787]: I1001 10:20:46.671846 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ptk7v" event={"ID":"9ccf3825-2cc8-491a-9a37-1973aa0dba30","Type":"ContainerDied","Data":"1ddb5f4d9acabf0d18435e45a5c896643f11d57078145a7b7edbd9fdf8048087"} Oct 01 10:20:46 crc kubenswrapper[4787]: I1001 10:20:46.671885 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ptk7v" event={"ID":"9ccf3825-2cc8-491a-9a37-1973aa0dba30","Type":"ContainerStarted","Data":"ad8d758a2715d1368f2c8e6b12527f2e933799e5c572a8bb396865559aea92c4"} Oct 01 10:20:46 crc kubenswrapper[4787]: I1001 10:20:46.686481 4787 generic.go:334] "Generic (PLEG): container finished" podID="dbd816a4-c3b2-4a94-9e2a-986a24503460" containerID="c6a51c4ea778de4b37d5b492ba9b671d93e81ac7988149c145a29c2259f6f8d3" exitCode=0 Oct 01 10:20:46 crc kubenswrapper[4787]: I1001 10:20:46.686529 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-blbn4" event={"ID":"dbd816a4-c3b2-4a94-9e2a-986a24503460","Type":"ContainerDied","Data":"c6a51c4ea778de4b37d5b492ba9b671d93e81ac7988149c145a29c2259f6f8d3"} Oct 01 10:20:46 crc kubenswrapper[4787]: E1001 10:20:46.776942 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9ccf3825_2cc8_491a_9a37_1973aa0dba30.slice/crio-1ddb5f4d9acabf0d18435e45a5c896643f11d57078145a7b7edbd9fdf8048087.scope\": RecentStats: unable to find data in memory cache]" Oct 01 10:20:47 crc kubenswrapper[4787]: I1001 10:20:47.700839 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-blbn4" event={"ID":"dbd816a4-c3b2-4a94-9e2a-986a24503460","Type":"ContainerStarted","Data":"3b211cb00c49ff6dc932c42d67ba96fca0650bfc0da2ea39e3e75ae5306fa999"} Oct 01 10:20:47 crc kubenswrapper[4787]: I1001 10:20:47.705578 4787 generic.go:334] "Generic (PLEG): container finished" podID="ce0e38fc-4300-4dd8-add8-4316579c9617" containerID="5de1454a3bb2a038c5ecca64f60772da9eda4f897ed3e0b0bd19c775b087dccf" exitCode=0 Oct 01 10:20:47 crc kubenswrapper[4787]: I1001 10:20:47.705626 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h4952" event={"ID":"ce0e38fc-4300-4dd8-add8-4316579c9617","Type":"ContainerDied","Data":"5de1454a3bb2a038c5ecca64f60772da9eda4f897ed3e0b0bd19c775b087dccf"} Oct 01 10:20:47 crc kubenswrapper[4787]: I1001 10:20:47.727282 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-blbn4" podStartSLOduration=2.188935195 podStartE2EDuration="4.727262814s" podCreationTimestamp="2025-10-01 10:20:43 +0000 UTC" firstStartedPulling="2025-10-01 10:20:44.63157484 +0000 UTC m=+2676.746718987" lastFinishedPulling="2025-10-01 10:20:47.169902449 +0000 UTC m=+2679.285046606" observedRunningTime="2025-10-01 10:20:47.719785783 +0000 UTC m=+2679.834929950" watchObservedRunningTime="2025-10-01 10:20:47.727262814 +0000 UTC m=+2679.842406971" Oct 01 10:20:48 crc kubenswrapper[4787]: I1001 10:20:48.722852 4787 generic.go:334] "Generic (PLEG): container finished" podID="9ccf3825-2cc8-491a-9a37-1973aa0dba30" containerID="13d8bdcb43d396b42164d312857ad9a1a7741455de428b7da822d7f3ceb49377" exitCode=0 Oct 01 10:20:48 crc kubenswrapper[4787]: I1001 10:20:48.722994 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ptk7v" event={"ID":"9ccf3825-2cc8-491a-9a37-1973aa0dba30","Type":"ContainerDied","Data":"13d8bdcb43d396b42164d312857ad9a1a7741455de428b7da822d7f3ceb49377"} Oct 01 10:20:49 crc kubenswrapper[4787]: I1001 10:20:49.734610 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ptk7v" event={"ID":"9ccf3825-2cc8-491a-9a37-1973aa0dba30","Type":"ContainerStarted","Data":"c465a337eec8bb67b2e98d91d81834d73cfa0c16d89a1af45bb999e746ba8893"} Oct 01 10:20:49 crc kubenswrapper[4787]: I1001 10:20:49.736869 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h4952" event={"ID":"ce0e38fc-4300-4dd8-add8-4316579c9617","Type":"ContainerStarted","Data":"162265bf3b4284cd3ef833e6fb1d799c9ee097238ae62ad0de773b82aaba7455"} Oct 01 10:20:49 crc kubenswrapper[4787]: I1001 10:20:49.761763 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ptk7v" podStartSLOduration=2.255592321 podStartE2EDuration="4.761748508s" podCreationTimestamp="2025-10-01 10:20:45 +0000 UTC" firstStartedPulling="2025-10-01 10:20:46.674106279 +0000 UTC m=+2678.789250436" lastFinishedPulling="2025-10-01 10:20:49.180262466 +0000 UTC m=+2681.295406623" observedRunningTime="2025-10-01 10:20:49.755231359 +0000 UTC m=+2681.870375536" watchObservedRunningTime="2025-10-01 10:20:49.761748508 +0000 UTC m=+2681.876892665" Oct 01 10:20:50 crc kubenswrapper[4787]: I1001 10:20:50.749507 4787 generic.go:334] "Generic (PLEG): container finished" podID="ce0e38fc-4300-4dd8-add8-4316579c9617" containerID="162265bf3b4284cd3ef833e6fb1d799c9ee097238ae62ad0de773b82aaba7455" exitCode=0 Oct 01 10:20:50 crc kubenswrapper[4787]: I1001 10:20:50.749595 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h4952" event={"ID":"ce0e38fc-4300-4dd8-add8-4316579c9617","Type":"ContainerDied","Data":"162265bf3b4284cd3ef833e6fb1d799c9ee097238ae62ad0de773b82aaba7455"} Oct 01 10:20:51 crc kubenswrapper[4787]: I1001 10:20:51.769848 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h4952" event={"ID":"ce0e38fc-4300-4dd8-add8-4316579c9617","Type":"ContainerStarted","Data":"8155e04c642a03a99e9667fb9b7f108a58bb4453b11e6ee8bef15be030c60907"} Oct 01 10:20:51 crc kubenswrapper[4787]: I1001 10:20:51.793114 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-h4952" podStartSLOduration=4.136776119 podStartE2EDuration="6.793092014s" podCreationTimestamp="2025-10-01 10:20:45 +0000 UTC" firstStartedPulling="2025-10-01 10:20:47.714979275 +0000 UTC m=+2679.830123432" lastFinishedPulling="2025-10-01 10:20:50.37129517 +0000 UTC m=+2682.486439327" observedRunningTime="2025-10-01 10:20:51.792257233 +0000 UTC m=+2683.907401400" watchObservedRunningTime="2025-10-01 10:20:51.793092014 +0000 UTC m=+2683.908236181" Oct 01 10:20:53 crc kubenswrapper[4787]: I1001 10:20:53.553598 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-blbn4" Oct 01 10:20:53 crc kubenswrapper[4787]: I1001 10:20:53.553983 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-blbn4" Oct 01 10:20:53 crc kubenswrapper[4787]: I1001 10:20:53.623663 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-blbn4" Oct 01 10:20:53 crc kubenswrapper[4787]: I1001 10:20:53.861006 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-blbn4" Oct 01 10:20:55 crc kubenswrapper[4787]: I1001 10:20:55.757445 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ptk7v" Oct 01 10:20:55 crc kubenswrapper[4787]: I1001 10:20:55.757775 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ptk7v" Oct 01 10:20:55 crc kubenswrapper[4787]: I1001 10:20:55.810365 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-blbn4"] Oct 01 10:20:55 crc kubenswrapper[4787]: I1001 10:20:55.824373 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-blbn4" podUID="dbd816a4-c3b2-4a94-9e2a-986a24503460" containerName="registry-server" containerID="cri-o://3b211cb00c49ff6dc932c42d67ba96fca0650bfc0da2ea39e3e75ae5306fa999" gracePeriod=2 Oct 01 10:20:55 crc kubenswrapper[4787]: I1001 10:20:55.943710 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-h4952" Oct 01 10:20:55 crc kubenswrapper[4787]: I1001 10:20:55.945194 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-h4952" Oct 01 10:20:56 crc kubenswrapper[4787]: I1001 10:20:56.007314 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-h4952" Oct 01 10:20:56 crc kubenswrapper[4787]: I1001 10:20:56.313387 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-blbn4" Oct 01 10:20:56 crc kubenswrapper[4787]: I1001 10:20:56.451404 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbd816a4-c3b2-4a94-9e2a-986a24503460-catalog-content\") pod \"dbd816a4-c3b2-4a94-9e2a-986a24503460\" (UID: \"dbd816a4-c3b2-4a94-9e2a-986a24503460\") " Oct 01 10:20:56 crc kubenswrapper[4787]: I1001 10:20:56.451474 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbd816a4-c3b2-4a94-9e2a-986a24503460-utilities\") pod \"dbd816a4-c3b2-4a94-9e2a-986a24503460\" (UID: \"dbd816a4-c3b2-4a94-9e2a-986a24503460\") " Oct 01 10:20:56 crc kubenswrapper[4787]: I1001 10:20:56.451590 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgr89\" (UniqueName: \"kubernetes.io/projected/dbd816a4-c3b2-4a94-9e2a-986a24503460-kube-api-access-hgr89\") pod \"dbd816a4-c3b2-4a94-9e2a-986a24503460\" (UID: \"dbd816a4-c3b2-4a94-9e2a-986a24503460\") " Oct 01 10:20:56 crc kubenswrapper[4787]: I1001 10:20:56.452187 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dbd816a4-c3b2-4a94-9e2a-986a24503460-utilities" (OuterVolumeSpecName: "utilities") pod "dbd816a4-c3b2-4a94-9e2a-986a24503460" (UID: "dbd816a4-c3b2-4a94-9e2a-986a24503460"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:20:56 crc kubenswrapper[4787]: I1001 10:20:56.458155 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbd816a4-c3b2-4a94-9e2a-986a24503460-kube-api-access-hgr89" (OuterVolumeSpecName: "kube-api-access-hgr89") pod "dbd816a4-c3b2-4a94-9e2a-986a24503460" (UID: "dbd816a4-c3b2-4a94-9e2a-986a24503460"). InnerVolumeSpecName "kube-api-access-hgr89". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:20:56 crc kubenswrapper[4787]: I1001 10:20:56.510489 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dbd816a4-c3b2-4a94-9e2a-986a24503460-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dbd816a4-c3b2-4a94-9e2a-986a24503460" (UID: "dbd816a4-c3b2-4a94-9e2a-986a24503460"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:20:56 crc kubenswrapper[4787]: I1001 10:20:56.554776 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgr89\" (UniqueName: \"kubernetes.io/projected/dbd816a4-c3b2-4a94-9e2a-986a24503460-kube-api-access-hgr89\") on node \"crc\" DevicePath \"\"" Oct 01 10:20:56 crc kubenswrapper[4787]: I1001 10:20:56.554816 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbd816a4-c3b2-4a94-9e2a-986a24503460-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 10:20:56 crc kubenswrapper[4787]: I1001 10:20:56.554831 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbd816a4-c3b2-4a94-9e2a-986a24503460-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 10:20:56 crc kubenswrapper[4787]: I1001 10:20:56.808802 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ptk7v" podUID="9ccf3825-2cc8-491a-9a37-1973aa0dba30" containerName="registry-server" probeResult="failure" output=< Oct 01 10:20:56 crc kubenswrapper[4787]: timeout: failed to connect service ":50051" within 1s Oct 01 10:20:56 crc kubenswrapper[4787]: > Oct 01 10:20:56 crc kubenswrapper[4787]: I1001 10:20:56.836031 4787 generic.go:334] "Generic (PLEG): container finished" podID="dbd816a4-c3b2-4a94-9e2a-986a24503460" containerID="3b211cb00c49ff6dc932c42d67ba96fca0650bfc0da2ea39e3e75ae5306fa999" exitCode=0 Oct 01 10:20:56 crc kubenswrapper[4787]: I1001 10:20:56.836131 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-blbn4" event={"ID":"dbd816a4-c3b2-4a94-9e2a-986a24503460","Type":"ContainerDied","Data":"3b211cb00c49ff6dc932c42d67ba96fca0650bfc0da2ea39e3e75ae5306fa999"} Oct 01 10:20:56 crc kubenswrapper[4787]: I1001 10:20:56.836197 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-blbn4" Oct 01 10:20:56 crc kubenswrapper[4787]: I1001 10:20:56.836240 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-blbn4" event={"ID":"dbd816a4-c3b2-4a94-9e2a-986a24503460","Type":"ContainerDied","Data":"275ca60dd3374b8357ef7a3d7e57a9297e82d1bfbdf0886cc3e6a2266291044e"} Oct 01 10:20:56 crc kubenswrapper[4787]: I1001 10:20:56.836268 4787 scope.go:117] "RemoveContainer" containerID="3b211cb00c49ff6dc932c42d67ba96fca0650bfc0da2ea39e3e75ae5306fa999" Oct 01 10:20:56 crc kubenswrapper[4787]: I1001 10:20:56.867425 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-blbn4"] Oct 01 10:20:56 crc kubenswrapper[4787]: I1001 10:20:56.874728 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-blbn4"] Oct 01 10:20:56 crc kubenswrapper[4787]: I1001 10:20:56.874752 4787 scope.go:117] "RemoveContainer" containerID="c6a51c4ea778de4b37d5b492ba9b671d93e81ac7988149c145a29c2259f6f8d3" Oct 01 10:20:56 crc kubenswrapper[4787]: I1001 10:20:56.906096 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-h4952" Oct 01 10:20:56 crc kubenswrapper[4787]: I1001 10:20:56.926104 4787 scope.go:117] "RemoveContainer" containerID="0dcb68f3cd52adb3a8624753edecac15d5b29859d4ab0b4b3ea07208f4d884f8" Oct 01 10:20:56 crc kubenswrapper[4787]: I1001 10:20:56.969521 4787 scope.go:117] "RemoveContainer" containerID="3b211cb00c49ff6dc932c42d67ba96fca0650bfc0da2ea39e3e75ae5306fa999" Oct 01 10:20:56 crc kubenswrapper[4787]: E1001 10:20:56.970540 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b211cb00c49ff6dc932c42d67ba96fca0650bfc0da2ea39e3e75ae5306fa999\": container with ID starting with 3b211cb00c49ff6dc932c42d67ba96fca0650bfc0da2ea39e3e75ae5306fa999 not found: ID does not exist" containerID="3b211cb00c49ff6dc932c42d67ba96fca0650bfc0da2ea39e3e75ae5306fa999" Oct 01 10:20:56 crc kubenswrapper[4787]: I1001 10:20:56.970617 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b211cb00c49ff6dc932c42d67ba96fca0650bfc0da2ea39e3e75ae5306fa999"} err="failed to get container status \"3b211cb00c49ff6dc932c42d67ba96fca0650bfc0da2ea39e3e75ae5306fa999\": rpc error: code = NotFound desc = could not find container \"3b211cb00c49ff6dc932c42d67ba96fca0650bfc0da2ea39e3e75ae5306fa999\": container with ID starting with 3b211cb00c49ff6dc932c42d67ba96fca0650bfc0da2ea39e3e75ae5306fa999 not found: ID does not exist" Oct 01 10:20:56 crc kubenswrapper[4787]: I1001 10:20:56.970673 4787 scope.go:117] "RemoveContainer" containerID="c6a51c4ea778de4b37d5b492ba9b671d93e81ac7988149c145a29c2259f6f8d3" Oct 01 10:20:56 crc kubenswrapper[4787]: E1001 10:20:56.975713 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6a51c4ea778de4b37d5b492ba9b671d93e81ac7988149c145a29c2259f6f8d3\": container with ID starting with c6a51c4ea778de4b37d5b492ba9b671d93e81ac7988149c145a29c2259f6f8d3 not found: ID does not exist" containerID="c6a51c4ea778de4b37d5b492ba9b671d93e81ac7988149c145a29c2259f6f8d3" Oct 01 10:20:56 crc kubenswrapper[4787]: I1001 10:20:56.975755 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6a51c4ea778de4b37d5b492ba9b671d93e81ac7988149c145a29c2259f6f8d3"} err="failed to get container status \"c6a51c4ea778de4b37d5b492ba9b671d93e81ac7988149c145a29c2259f6f8d3\": rpc error: code = NotFound desc = could not find container \"c6a51c4ea778de4b37d5b492ba9b671d93e81ac7988149c145a29c2259f6f8d3\": container with ID starting with c6a51c4ea778de4b37d5b492ba9b671d93e81ac7988149c145a29c2259f6f8d3 not found: ID does not exist" Oct 01 10:20:56 crc kubenswrapper[4787]: I1001 10:20:56.975782 4787 scope.go:117] "RemoveContainer" containerID="0dcb68f3cd52adb3a8624753edecac15d5b29859d4ab0b4b3ea07208f4d884f8" Oct 01 10:20:56 crc kubenswrapper[4787]: E1001 10:20:56.976305 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0dcb68f3cd52adb3a8624753edecac15d5b29859d4ab0b4b3ea07208f4d884f8\": container with ID starting with 0dcb68f3cd52adb3a8624753edecac15d5b29859d4ab0b4b3ea07208f4d884f8 not found: ID does not exist" containerID="0dcb68f3cd52adb3a8624753edecac15d5b29859d4ab0b4b3ea07208f4d884f8" Oct 01 10:20:56 crc kubenswrapper[4787]: I1001 10:20:56.976359 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0dcb68f3cd52adb3a8624753edecac15d5b29859d4ab0b4b3ea07208f4d884f8"} err="failed to get container status \"0dcb68f3cd52adb3a8624753edecac15d5b29859d4ab0b4b3ea07208f4d884f8\": rpc error: code = NotFound desc = could not find container \"0dcb68f3cd52adb3a8624753edecac15d5b29859d4ab0b4b3ea07208f4d884f8\": container with ID starting with 0dcb68f3cd52adb3a8624753edecac15d5b29859d4ab0b4b3ea07208f4d884f8 not found: ID does not exist" Oct 01 10:20:58 crc kubenswrapper[4787]: I1001 10:20:58.407946 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-h4952"] Oct 01 10:20:58 crc kubenswrapper[4787]: I1001 10:20:58.537445 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbd816a4-c3b2-4a94-9e2a-986a24503460" path="/var/lib/kubelet/pods/dbd816a4-c3b2-4a94-9e2a-986a24503460/volumes" Oct 01 10:20:59 crc kubenswrapper[4787]: I1001 10:20:59.872814 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-h4952" podUID="ce0e38fc-4300-4dd8-add8-4316579c9617" containerName="registry-server" containerID="cri-o://8155e04c642a03a99e9667fb9b7f108a58bb4453b11e6ee8bef15be030c60907" gracePeriod=2 Oct 01 10:21:00 crc kubenswrapper[4787]: I1001 10:21:00.418103 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h4952" Oct 01 10:21:00 crc kubenswrapper[4787]: I1001 10:21:00.432015 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce0e38fc-4300-4dd8-add8-4316579c9617-catalog-content\") pod \"ce0e38fc-4300-4dd8-add8-4316579c9617\" (UID: \"ce0e38fc-4300-4dd8-add8-4316579c9617\") " Oct 01 10:21:00 crc kubenswrapper[4787]: I1001 10:21:00.432117 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vsjjs\" (UniqueName: \"kubernetes.io/projected/ce0e38fc-4300-4dd8-add8-4316579c9617-kube-api-access-vsjjs\") pod \"ce0e38fc-4300-4dd8-add8-4316579c9617\" (UID: \"ce0e38fc-4300-4dd8-add8-4316579c9617\") " Oct 01 10:21:00 crc kubenswrapper[4787]: I1001 10:21:00.432248 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce0e38fc-4300-4dd8-add8-4316579c9617-utilities\") pod \"ce0e38fc-4300-4dd8-add8-4316579c9617\" (UID: \"ce0e38fc-4300-4dd8-add8-4316579c9617\") " Oct 01 10:21:00 crc kubenswrapper[4787]: I1001 10:21:00.433825 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce0e38fc-4300-4dd8-add8-4316579c9617-utilities" (OuterVolumeSpecName: "utilities") pod "ce0e38fc-4300-4dd8-add8-4316579c9617" (UID: "ce0e38fc-4300-4dd8-add8-4316579c9617"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:21:00 crc kubenswrapper[4787]: I1001 10:21:00.441244 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce0e38fc-4300-4dd8-add8-4316579c9617-kube-api-access-vsjjs" (OuterVolumeSpecName: "kube-api-access-vsjjs") pod "ce0e38fc-4300-4dd8-add8-4316579c9617" (UID: "ce0e38fc-4300-4dd8-add8-4316579c9617"). InnerVolumeSpecName "kube-api-access-vsjjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:21:00 crc kubenswrapper[4787]: I1001 10:21:00.502556 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce0e38fc-4300-4dd8-add8-4316579c9617-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ce0e38fc-4300-4dd8-add8-4316579c9617" (UID: "ce0e38fc-4300-4dd8-add8-4316579c9617"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:21:00 crc kubenswrapper[4787]: I1001 10:21:00.534839 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce0e38fc-4300-4dd8-add8-4316579c9617-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 10:21:00 crc kubenswrapper[4787]: I1001 10:21:00.534867 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vsjjs\" (UniqueName: \"kubernetes.io/projected/ce0e38fc-4300-4dd8-add8-4316579c9617-kube-api-access-vsjjs\") on node \"crc\" DevicePath \"\"" Oct 01 10:21:00 crc kubenswrapper[4787]: I1001 10:21:00.534906 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce0e38fc-4300-4dd8-add8-4316579c9617-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 10:21:00 crc kubenswrapper[4787]: I1001 10:21:00.888291 4787 generic.go:334] "Generic (PLEG): container finished" podID="ce0e38fc-4300-4dd8-add8-4316579c9617" containerID="8155e04c642a03a99e9667fb9b7f108a58bb4453b11e6ee8bef15be030c60907" exitCode=0 Oct 01 10:21:00 crc kubenswrapper[4787]: I1001 10:21:00.888348 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h4952" Oct 01 10:21:00 crc kubenswrapper[4787]: I1001 10:21:00.888376 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h4952" event={"ID":"ce0e38fc-4300-4dd8-add8-4316579c9617","Type":"ContainerDied","Data":"8155e04c642a03a99e9667fb9b7f108a58bb4453b11e6ee8bef15be030c60907"} Oct 01 10:21:00 crc kubenswrapper[4787]: I1001 10:21:00.888879 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h4952" event={"ID":"ce0e38fc-4300-4dd8-add8-4316579c9617","Type":"ContainerDied","Data":"3f41a7b5f466f643ee3ea454959dec54cb68be7dc5312a32803876d5c24f09b9"} Oct 01 10:21:00 crc kubenswrapper[4787]: I1001 10:21:00.888910 4787 scope.go:117] "RemoveContainer" containerID="8155e04c642a03a99e9667fb9b7f108a58bb4453b11e6ee8bef15be030c60907" Oct 01 10:21:00 crc kubenswrapper[4787]: I1001 10:21:00.913503 4787 scope.go:117] "RemoveContainer" containerID="162265bf3b4284cd3ef833e6fb1d799c9ee097238ae62ad0de773b82aaba7455" Oct 01 10:21:00 crc kubenswrapper[4787]: I1001 10:21:00.919903 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-h4952"] Oct 01 10:21:00 crc kubenswrapper[4787]: I1001 10:21:00.927756 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-h4952"] Oct 01 10:21:00 crc kubenswrapper[4787]: I1001 10:21:00.961723 4787 scope.go:117] "RemoveContainer" containerID="5de1454a3bb2a038c5ecca64f60772da9eda4f897ed3e0b0bd19c775b087dccf" Oct 01 10:21:00 crc kubenswrapper[4787]: I1001 10:21:00.985428 4787 scope.go:117] "RemoveContainer" containerID="8155e04c642a03a99e9667fb9b7f108a58bb4453b11e6ee8bef15be030c60907" Oct 01 10:21:00 crc kubenswrapper[4787]: E1001 10:21:00.985852 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8155e04c642a03a99e9667fb9b7f108a58bb4453b11e6ee8bef15be030c60907\": container with ID starting with 8155e04c642a03a99e9667fb9b7f108a58bb4453b11e6ee8bef15be030c60907 not found: ID does not exist" containerID="8155e04c642a03a99e9667fb9b7f108a58bb4453b11e6ee8bef15be030c60907" Oct 01 10:21:00 crc kubenswrapper[4787]: I1001 10:21:00.985888 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8155e04c642a03a99e9667fb9b7f108a58bb4453b11e6ee8bef15be030c60907"} err="failed to get container status \"8155e04c642a03a99e9667fb9b7f108a58bb4453b11e6ee8bef15be030c60907\": rpc error: code = NotFound desc = could not find container \"8155e04c642a03a99e9667fb9b7f108a58bb4453b11e6ee8bef15be030c60907\": container with ID starting with 8155e04c642a03a99e9667fb9b7f108a58bb4453b11e6ee8bef15be030c60907 not found: ID does not exist" Oct 01 10:21:00 crc kubenswrapper[4787]: I1001 10:21:00.985909 4787 scope.go:117] "RemoveContainer" containerID="162265bf3b4284cd3ef833e6fb1d799c9ee097238ae62ad0de773b82aaba7455" Oct 01 10:21:00 crc kubenswrapper[4787]: E1001 10:21:00.986466 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"162265bf3b4284cd3ef833e6fb1d799c9ee097238ae62ad0de773b82aaba7455\": container with ID starting with 162265bf3b4284cd3ef833e6fb1d799c9ee097238ae62ad0de773b82aaba7455 not found: ID does not exist" containerID="162265bf3b4284cd3ef833e6fb1d799c9ee097238ae62ad0de773b82aaba7455" Oct 01 10:21:00 crc kubenswrapper[4787]: I1001 10:21:00.986494 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"162265bf3b4284cd3ef833e6fb1d799c9ee097238ae62ad0de773b82aaba7455"} err="failed to get container status \"162265bf3b4284cd3ef833e6fb1d799c9ee097238ae62ad0de773b82aaba7455\": rpc error: code = NotFound desc = could not find container \"162265bf3b4284cd3ef833e6fb1d799c9ee097238ae62ad0de773b82aaba7455\": container with ID starting with 162265bf3b4284cd3ef833e6fb1d799c9ee097238ae62ad0de773b82aaba7455 not found: ID does not exist" Oct 01 10:21:00 crc kubenswrapper[4787]: I1001 10:21:00.986512 4787 scope.go:117] "RemoveContainer" containerID="5de1454a3bb2a038c5ecca64f60772da9eda4f897ed3e0b0bd19c775b087dccf" Oct 01 10:21:00 crc kubenswrapper[4787]: E1001 10:21:00.986834 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5de1454a3bb2a038c5ecca64f60772da9eda4f897ed3e0b0bd19c775b087dccf\": container with ID starting with 5de1454a3bb2a038c5ecca64f60772da9eda4f897ed3e0b0bd19c775b087dccf not found: ID does not exist" containerID="5de1454a3bb2a038c5ecca64f60772da9eda4f897ed3e0b0bd19c775b087dccf" Oct 01 10:21:00 crc kubenswrapper[4787]: I1001 10:21:00.986857 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5de1454a3bb2a038c5ecca64f60772da9eda4f897ed3e0b0bd19c775b087dccf"} err="failed to get container status \"5de1454a3bb2a038c5ecca64f60772da9eda4f897ed3e0b0bd19c775b087dccf\": rpc error: code = NotFound desc = could not find container \"5de1454a3bb2a038c5ecca64f60772da9eda4f897ed3e0b0bd19c775b087dccf\": container with ID starting with 5de1454a3bb2a038c5ecca64f60772da9eda4f897ed3e0b0bd19c775b087dccf not found: ID does not exist" Oct 01 10:21:02 crc kubenswrapper[4787]: I1001 10:21:02.533471 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce0e38fc-4300-4dd8-add8-4316579c9617" path="/var/lib/kubelet/pods/ce0e38fc-4300-4dd8-add8-4316579c9617/volumes" Oct 01 10:21:05 crc kubenswrapper[4787]: I1001 10:21:05.812039 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ptk7v" Oct 01 10:21:05 crc kubenswrapper[4787]: I1001 10:21:05.868984 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ptk7v" Oct 01 10:21:06 crc kubenswrapper[4787]: I1001 10:21:06.051313 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ptk7v"] Oct 01 10:21:06 crc kubenswrapper[4787]: I1001 10:21:06.978321 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ptk7v" podUID="9ccf3825-2cc8-491a-9a37-1973aa0dba30" containerName="registry-server" containerID="cri-o://c465a337eec8bb67b2e98d91d81834d73cfa0c16d89a1af45bb999e746ba8893" gracePeriod=2 Oct 01 10:21:07 crc kubenswrapper[4787]: I1001 10:21:07.506740 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ptk7v" Oct 01 10:21:07 crc kubenswrapper[4787]: I1001 10:21:07.673162 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffptl\" (UniqueName: \"kubernetes.io/projected/9ccf3825-2cc8-491a-9a37-1973aa0dba30-kube-api-access-ffptl\") pod \"9ccf3825-2cc8-491a-9a37-1973aa0dba30\" (UID: \"9ccf3825-2cc8-491a-9a37-1973aa0dba30\") " Oct 01 10:21:07 crc kubenswrapper[4787]: I1001 10:21:07.673235 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ccf3825-2cc8-491a-9a37-1973aa0dba30-utilities\") pod \"9ccf3825-2cc8-491a-9a37-1973aa0dba30\" (UID: \"9ccf3825-2cc8-491a-9a37-1973aa0dba30\") " Oct 01 10:21:07 crc kubenswrapper[4787]: I1001 10:21:07.673730 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ccf3825-2cc8-491a-9a37-1973aa0dba30-catalog-content\") pod \"9ccf3825-2cc8-491a-9a37-1973aa0dba30\" (UID: \"9ccf3825-2cc8-491a-9a37-1973aa0dba30\") " Oct 01 10:21:07 crc kubenswrapper[4787]: I1001 10:21:07.674101 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ccf3825-2cc8-491a-9a37-1973aa0dba30-utilities" (OuterVolumeSpecName: "utilities") pod "9ccf3825-2cc8-491a-9a37-1973aa0dba30" (UID: "9ccf3825-2cc8-491a-9a37-1973aa0dba30"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:21:07 crc kubenswrapper[4787]: I1001 10:21:07.674668 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ccf3825-2cc8-491a-9a37-1973aa0dba30-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 10:21:07 crc kubenswrapper[4787]: I1001 10:21:07.679463 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ccf3825-2cc8-491a-9a37-1973aa0dba30-kube-api-access-ffptl" (OuterVolumeSpecName: "kube-api-access-ffptl") pod "9ccf3825-2cc8-491a-9a37-1973aa0dba30" (UID: "9ccf3825-2cc8-491a-9a37-1973aa0dba30"). InnerVolumeSpecName "kube-api-access-ffptl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:21:07 crc kubenswrapper[4787]: I1001 10:21:07.751228 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ccf3825-2cc8-491a-9a37-1973aa0dba30-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9ccf3825-2cc8-491a-9a37-1973aa0dba30" (UID: "9ccf3825-2cc8-491a-9a37-1973aa0dba30"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:21:07 crc kubenswrapper[4787]: I1001 10:21:07.776816 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffptl\" (UniqueName: \"kubernetes.io/projected/9ccf3825-2cc8-491a-9a37-1973aa0dba30-kube-api-access-ffptl\") on node \"crc\" DevicePath \"\"" Oct 01 10:21:07 crc kubenswrapper[4787]: I1001 10:21:07.776857 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ccf3825-2cc8-491a-9a37-1973aa0dba30-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 10:21:07 crc kubenswrapper[4787]: I1001 10:21:07.987449 4787 generic.go:334] "Generic (PLEG): container finished" podID="9ccf3825-2cc8-491a-9a37-1973aa0dba30" containerID="c465a337eec8bb67b2e98d91d81834d73cfa0c16d89a1af45bb999e746ba8893" exitCode=0 Oct 01 10:21:07 crc kubenswrapper[4787]: I1001 10:21:07.987503 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ptk7v" Oct 01 10:21:07 crc kubenswrapper[4787]: I1001 10:21:07.987523 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ptk7v" event={"ID":"9ccf3825-2cc8-491a-9a37-1973aa0dba30","Type":"ContainerDied","Data":"c465a337eec8bb67b2e98d91d81834d73cfa0c16d89a1af45bb999e746ba8893"} Oct 01 10:21:07 crc kubenswrapper[4787]: I1001 10:21:07.989261 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ptk7v" event={"ID":"9ccf3825-2cc8-491a-9a37-1973aa0dba30","Type":"ContainerDied","Data":"ad8d758a2715d1368f2c8e6b12527f2e933799e5c572a8bb396865559aea92c4"} Oct 01 10:21:07 crc kubenswrapper[4787]: I1001 10:21:07.989284 4787 scope.go:117] "RemoveContainer" containerID="c465a337eec8bb67b2e98d91d81834d73cfa0c16d89a1af45bb999e746ba8893" Oct 01 10:21:08 crc kubenswrapper[4787]: I1001 10:21:08.022783 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ptk7v"] Oct 01 10:21:08 crc kubenswrapper[4787]: I1001 10:21:08.027637 4787 scope.go:117] "RemoveContainer" containerID="13d8bdcb43d396b42164d312857ad9a1a7741455de428b7da822d7f3ceb49377" Oct 01 10:21:08 crc kubenswrapper[4787]: I1001 10:21:08.034603 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ptk7v"] Oct 01 10:21:08 crc kubenswrapper[4787]: I1001 10:21:08.055399 4787 scope.go:117] "RemoveContainer" containerID="1ddb5f4d9acabf0d18435e45a5c896643f11d57078145a7b7edbd9fdf8048087" Oct 01 10:21:08 crc kubenswrapper[4787]: I1001 10:21:08.098709 4787 scope.go:117] "RemoveContainer" containerID="c465a337eec8bb67b2e98d91d81834d73cfa0c16d89a1af45bb999e746ba8893" Oct 01 10:21:08 crc kubenswrapper[4787]: E1001 10:21:08.099186 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c465a337eec8bb67b2e98d91d81834d73cfa0c16d89a1af45bb999e746ba8893\": container with ID starting with c465a337eec8bb67b2e98d91d81834d73cfa0c16d89a1af45bb999e746ba8893 not found: ID does not exist" containerID="c465a337eec8bb67b2e98d91d81834d73cfa0c16d89a1af45bb999e746ba8893" Oct 01 10:21:08 crc kubenswrapper[4787]: I1001 10:21:08.099231 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c465a337eec8bb67b2e98d91d81834d73cfa0c16d89a1af45bb999e746ba8893"} err="failed to get container status \"c465a337eec8bb67b2e98d91d81834d73cfa0c16d89a1af45bb999e746ba8893\": rpc error: code = NotFound desc = could not find container \"c465a337eec8bb67b2e98d91d81834d73cfa0c16d89a1af45bb999e746ba8893\": container with ID starting with c465a337eec8bb67b2e98d91d81834d73cfa0c16d89a1af45bb999e746ba8893 not found: ID does not exist" Oct 01 10:21:08 crc kubenswrapper[4787]: I1001 10:21:08.099257 4787 scope.go:117] "RemoveContainer" containerID="13d8bdcb43d396b42164d312857ad9a1a7741455de428b7da822d7f3ceb49377" Oct 01 10:21:08 crc kubenswrapper[4787]: E1001 10:21:08.099670 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13d8bdcb43d396b42164d312857ad9a1a7741455de428b7da822d7f3ceb49377\": container with ID starting with 13d8bdcb43d396b42164d312857ad9a1a7741455de428b7da822d7f3ceb49377 not found: ID does not exist" containerID="13d8bdcb43d396b42164d312857ad9a1a7741455de428b7da822d7f3ceb49377" Oct 01 10:21:08 crc kubenswrapper[4787]: I1001 10:21:08.099697 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13d8bdcb43d396b42164d312857ad9a1a7741455de428b7da822d7f3ceb49377"} err="failed to get container status \"13d8bdcb43d396b42164d312857ad9a1a7741455de428b7da822d7f3ceb49377\": rpc error: code = NotFound desc = could not find container \"13d8bdcb43d396b42164d312857ad9a1a7741455de428b7da822d7f3ceb49377\": container with ID starting with 13d8bdcb43d396b42164d312857ad9a1a7741455de428b7da822d7f3ceb49377 not found: ID does not exist" Oct 01 10:21:08 crc kubenswrapper[4787]: I1001 10:21:08.099715 4787 scope.go:117] "RemoveContainer" containerID="1ddb5f4d9acabf0d18435e45a5c896643f11d57078145a7b7edbd9fdf8048087" Oct 01 10:21:08 crc kubenswrapper[4787]: E1001 10:21:08.099973 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ddb5f4d9acabf0d18435e45a5c896643f11d57078145a7b7edbd9fdf8048087\": container with ID starting with 1ddb5f4d9acabf0d18435e45a5c896643f11d57078145a7b7edbd9fdf8048087 not found: ID does not exist" containerID="1ddb5f4d9acabf0d18435e45a5c896643f11d57078145a7b7edbd9fdf8048087" Oct 01 10:21:08 crc kubenswrapper[4787]: I1001 10:21:08.099997 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ddb5f4d9acabf0d18435e45a5c896643f11d57078145a7b7edbd9fdf8048087"} err="failed to get container status \"1ddb5f4d9acabf0d18435e45a5c896643f11d57078145a7b7edbd9fdf8048087\": rpc error: code = NotFound desc = could not find container \"1ddb5f4d9acabf0d18435e45a5c896643f11d57078145a7b7edbd9fdf8048087\": container with ID starting with 1ddb5f4d9acabf0d18435e45a5c896643f11d57078145a7b7edbd9fdf8048087 not found: ID does not exist" Oct 01 10:21:08 crc kubenswrapper[4787]: I1001 10:21:08.538554 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ccf3825-2cc8-491a-9a37-1973aa0dba30" path="/var/lib/kubelet/pods/9ccf3825-2cc8-491a-9a37-1973aa0dba30/volumes" Oct 01 10:22:24 crc kubenswrapper[4787]: I1001 10:22:24.578596 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-w4jbs"] Oct 01 10:22:24 crc kubenswrapper[4787]: E1001 10:22:24.579607 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce0e38fc-4300-4dd8-add8-4316579c9617" containerName="extract-content" Oct 01 10:22:24 crc kubenswrapper[4787]: I1001 10:22:24.579624 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce0e38fc-4300-4dd8-add8-4316579c9617" containerName="extract-content" Oct 01 10:22:24 crc kubenswrapper[4787]: E1001 10:22:24.579643 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ccf3825-2cc8-491a-9a37-1973aa0dba30" containerName="extract-content" Oct 01 10:22:24 crc kubenswrapper[4787]: I1001 10:22:24.579649 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ccf3825-2cc8-491a-9a37-1973aa0dba30" containerName="extract-content" Oct 01 10:22:24 crc kubenswrapper[4787]: E1001 10:22:24.579661 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbd816a4-c3b2-4a94-9e2a-986a24503460" containerName="extract-utilities" Oct 01 10:22:24 crc kubenswrapper[4787]: I1001 10:22:24.579667 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbd816a4-c3b2-4a94-9e2a-986a24503460" containerName="extract-utilities" Oct 01 10:22:24 crc kubenswrapper[4787]: E1001 10:22:24.579690 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbd816a4-c3b2-4a94-9e2a-986a24503460" containerName="extract-content" Oct 01 10:22:24 crc kubenswrapper[4787]: I1001 10:22:24.579698 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbd816a4-c3b2-4a94-9e2a-986a24503460" containerName="extract-content" Oct 01 10:22:24 crc kubenswrapper[4787]: E1001 10:22:24.579711 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce0e38fc-4300-4dd8-add8-4316579c9617" containerName="registry-server" Oct 01 10:22:24 crc kubenswrapper[4787]: I1001 10:22:24.579719 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce0e38fc-4300-4dd8-add8-4316579c9617" containerName="registry-server" Oct 01 10:22:24 crc kubenswrapper[4787]: E1001 10:22:24.579734 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ccf3825-2cc8-491a-9a37-1973aa0dba30" containerName="extract-utilities" Oct 01 10:22:24 crc kubenswrapper[4787]: I1001 10:22:24.579740 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ccf3825-2cc8-491a-9a37-1973aa0dba30" containerName="extract-utilities" Oct 01 10:22:24 crc kubenswrapper[4787]: E1001 10:22:24.579754 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ccf3825-2cc8-491a-9a37-1973aa0dba30" containerName="registry-server" Oct 01 10:22:24 crc kubenswrapper[4787]: I1001 10:22:24.579760 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ccf3825-2cc8-491a-9a37-1973aa0dba30" containerName="registry-server" Oct 01 10:22:24 crc kubenswrapper[4787]: E1001 10:22:24.579774 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbd816a4-c3b2-4a94-9e2a-986a24503460" containerName="registry-server" Oct 01 10:22:24 crc kubenswrapper[4787]: I1001 10:22:24.579781 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbd816a4-c3b2-4a94-9e2a-986a24503460" containerName="registry-server" Oct 01 10:22:24 crc kubenswrapper[4787]: E1001 10:22:24.579797 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce0e38fc-4300-4dd8-add8-4316579c9617" containerName="extract-utilities" Oct 01 10:22:24 crc kubenswrapper[4787]: I1001 10:22:24.579805 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce0e38fc-4300-4dd8-add8-4316579c9617" containerName="extract-utilities" Oct 01 10:22:24 crc kubenswrapper[4787]: I1001 10:22:24.580021 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ccf3825-2cc8-491a-9a37-1973aa0dba30" containerName="registry-server" Oct 01 10:22:24 crc kubenswrapper[4787]: I1001 10:22:24.580033 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbd816a4-c3b2-4a94-9e2a-986a24503460" containerName="registry-server" Oct 01 10:22:24 crc kubenswrapper[4787]: I1001 10:22:24.580050 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce0e38fc-4300-4dd8-add8-4316579c9617" containerName="registry-server" Oct 01 10:22:24 crc kubenswrapper[4787]: I1001 10:22:24.581631 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w4jbs" Oct 01 10:22:24 crc kubenswrapper[4787]: I1001 10:22:24.589204 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w4jbs"] Oct 01 10:22:24 crc kubenswrapper[4787]: I1001 10:22:24.643340 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rnkn\" (UniqueName: \"kubernetes.io/projected/bc3deca0-6cd1-43f6-84f7-98b600eaa4e7-kube-api-access-2rnkn\") pod \"redhat-marketplace-w4jbs\" (UID: \"bc3deca0-6cd1-43f6-84f7-98b600eaa4e7\") " pod="openshift-marketplace/redhat-marketplace-w4jbs" Oct 01 10:22:24 crc kubenswrapper[4787]: I1001 10:22:24.643681 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc3deca0-6cd1-43f6-84f7-98b600eaa4e7-utilities\") pod \"redhat-marketplace-w4jbs\" (UID: \"bc3deca0-6cd1-43f6-84f7-98b600eaa4e7\") " pod="openshift-marketplace/redhat-marketplace-w4jbs" Oct 01 10:22:24 crc kubenswrapper[4787]: I1001 10:22:24.643828 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc3deca0-6cd1-43f6-84f7-98b600eaa4e7-catalog-content\") pod \"redhat-marketplace-w4jbs\" (UID: \"bc3deca0-6cd1-43f6-84f7-98b600eaa4e7\") " pod="openshift-marketplace/redhat-marketplace-w4jbs" Oct 01 10:22:24 crc kubenswrapper[4787]: I1001 10:22:24.745334 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc3deca0-6cd1-43f6-84f7-98b600eaa4e7-utilities\") pod \"redhat-marketplace-w4jbs\" (UID: \"bc3deca0-6cd1-43f6-84f7-98b600eaa4e7\") " pod="openshift-marketplace/redhat-marketplace-w4jbs" Oct 01 10:22:24 crc kubenswrapper[4787]: I1001 10:22:24.745454 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc3deca0-6cd1-43f6-84f7-98b600eaa4e7-catalog-content\") pod \"redhat-marketplace-w4jbs\" (UID: \"bc3deca0-6cd1-43f6-84f7-98b600eaa4e7\") " pod="openshift-marketplace/redhat-marketplace-w4jbs" Oct 01 10:22:24 crc kubenswrapper[4787]: I1001 10:22:24.745508 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rnkn\" (UniqueName: \"kubernetes.io/projected/bc3deca0-6cd1-43f6-84f7-98b600eaa4e7-kube-api-access-2rnkn\") pod \"redhat-marketplace-w4jbs\" (UID: \"bc3deca0-6cd1-43f6-84f7-98b600eaa4e7\") " pod="openshift-marketplace/redhat-marketplace-w4jbs" Oct 01 10:22:24 crc kubenswrapper[4787]: I1001 10:22:24.746522 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc3deca0-6cd1-43f6-84f7-98b600eaa4e7-utilities\") pod \"redhat-marketplace-w4jbs\" (UID: \"bc3deca0-6cd1-43f6-84f7-98b600eaa4e7\") " pod="openshift-marketplace/redhat-marketplace-w4jbs" Oct 01 10:22:24 crc kubenswrapper[4787]: I1001 10:22:24.746542 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc3deca0-6cd1-43f6-84f7-98b600eaa4e7-catalog-content\") pod \"redhat-marketplace-w4jbs\" (UID: \"bc3deca0-6cd1-43f6-84f7-98b600eaa4e7\") " pod="openshift-marketplace/redhat-marketplace-w4jbs" Oct 01 10:22:24 crc kubenswrapper[4787]: I1001 10:22:24.769914 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rnkn\" (UniqueName: \"kubernetes.io/projected/bc3deca0-6cd1-43f6-84f7-98b600eaa4e7-kube-api-access-2rnkn\") pod \"redhat-marketplace-w4jbs\" (UID: \"bc3deca0-6cd1-43f6-84f7-98b600eaa4e7\") " pod="openshift-marketplace/redhat-marketplace-w4jbs" Oct 01 10:22:24 crc kubenswrapper[4787]: I1001 10:22:24.938273 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w4jbs" Oct 01 10:22:25 crc kubenswrapper[4787]: I1001 10:22:25.400035 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w4jbs"] Oct 01 10:22:25 crc kubenswrapper[4787]: I1001 10:22:25.813960 4787 generic.go:334] "Generic (PLEG): container finished" podID="bc3deca0-6cd1-43f6-84f7-98b600eaa4e7" containerID="919d5e817538ec492a53fe5c0c7bc7eeb5fe1c706fc140027eea45ee5fe106af" exitCode=0 Oct 01 10:22:25 crc kubenswrapper[4787]: I1001 10:22:25.814434 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w4jbs" event={"ID":"bc3deca0-6cd1-43f6-84f7-98b600eaa4e7","Type":"ContainerDied","Data":"919d5e817538ec492a53fe5c0c7bc7eeb5fe1c706fc140027eea45ee5fe106af"} Oct 01 10:22:25 crc kubenswrapper[4787]: I1001 10:22:25.814467 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w4jbs" event={"ID":"bc3deca0-6cd1-43f6-84f7-98b600eaa4e7","Type":"ContainerStarted","Data":"40cb4b049d239567ffe6a88066bd5e7dc4772a7b71421936bde81299b729f7d0"} Oct 01 10:22:26 crc kubenswrapper[4787]: I1001 10:22:26.824492 4787 generic.go:334] "Generic (PLEG): container finished" podID="bc3deca0-6cd1-43f6-84f7-98b600eaa4e7" containerID="860b7cb0b20f12a2ac5b3022980865dc8f97f5da1930da8f6ab0c2c9cfd344d5" exitCode=0 Oct 01 10:22:26 crc kubenswrapper[4787]: I1001 10:22:26.824608 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w4jbs" event={"ID":"bc3deca0-6cd1-43f6-84f7-98b600eaa4e7","Type":"ContainerDied","Data":"860b7cb0b20f12a2ac5b3022980865dc8f97f5da1930da8f6ab0c2c9cfd344d5"} Oct 01 10:22:27 crc kubenswrapper[4787]: I1001 10:22:27.835221 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w4jbs" event={"ID":"bc3deca0-6cd1-43f6-84f7-98b600eaa4e7","Type":"ContainerStarted","Data":"4387d1306a4a3ef35860c3b5e8dece9d951d3dfa943369c82d3242074abc3cea"} Oct 01 10:22:27 crc kubenswrapper[4787]: I1001 10:22:27.860809 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-w4jbs" podStartSLOduration=2.381020944 podStartE2EDuration="3.860793637s" podCreationTimestamp="2025-10-01 10:22:24 +0000 UTC" firstStartedPulling="2025-10-01 10:22:25.817889009 +0000 UTC m=+2777.933033166" lastFinishedPulling="2025-10-01 10:22:27.297661702 +0000 UTC m=+2779.412805859" observedRunningTime="2025-10-01 10:22:27.855768565 +0000 UTC m=+2779.970912722" watchObservedRunningTime="2025-10-01 10:22:27.860793637 +0000 UTC m=+2779.975937794" Oct 01 10:22:34 crc kubenswrapper[4787]: I1001 10:22:34.938886 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-w4jbs" Oct 01 10:22:34 crc kubenswrapper[4787]: I1001 10:22:34.940916 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-w4jbs" Oct 01 10:22:34 crc kubenswrapper[4787]: I1001 10:22:34.994347 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-w4jbs" Oct 01 10:22:35 crc kubenswrapper[4787]: I1001 10:22:35.950938 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-w4jbs" Oct 01 10:22:35 crc kubenswrapper[4787]: I1001 10:22:35.993831 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w4jbs"] Oct 01 10:22:37 crc kubenswrapper[4787]: I1001 10:22:37.922700 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-w4jbs" podUID="bc3deca0-6cd1-43f6-84f7-98b600eaa4e7" containerName="registry-server" containerID="cri-o://4387d1306a4a3ef35860c3b5e8dece9d951d3dfa943369c82d3242074abc3cea" gracePeriod=2 Oct 01 10:22:38 crc kubenswrapper[4787]: I1001 10:22:38.431949 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w4jbs" Oct 01 10:22:38 crc kubenswrapper[4787]: I1001 10:22:38.554089 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc3deca0-6cd1-43f6-84f7-98b600eaa4e7-utilities\") pod \"bc3deca0-6cd1-43f6-84f7-98b600eaa4e7\" (UID: \"bc3deca0-6cd1-43f6-84f7-98b600eaa4e7\") " Oct 01 10:22:38 crc kubenswrapper[4787]: I1001 10:22:38.554541 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc3deca0-6cd1-43f6-84f7-98b600eaa4e7-catalog-content\") pod \"bc3deca0-6cd1-43f6-84f7-98b600eaa4e7\" (UID: \"bc3deca0-6cd1-43f6-84f7-98b600eaa4e7\") " Oct 01 10:22:38 crc kubenswrapper[4787]: I1001 10:22:38.554584 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2rnkn\" (UniqueName: \"kubernetes.io/projected/bc3deca0-6cd1-43f6-84f7-98b600eaa4e7-kube-api-access-2rnkn\") pod \"bc3deca0-6cd1-43f6-84f7-98b600eaa4e7\" (UID: \"bc3deca0-6cd1-43f6-84f7-98b600eaa4e7\") " Oct 01 10:22:38 crc kubenswrapper[4787]: I1001 10:22:38.554956 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc3deca0-6cd1-43f6-84f7-98b600eaa4e7-utilities" (OuterVolumeSpecName: "utilities") pod "bc3deca0-6cd1-43f6-84f7-98b600eaa4e7" (UID: "bc3deca0-6cd1-43f6-84f7-98b600eaa4e7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:22:38 crc kubenswrapper[4787]: I1001 10:22:38.555662 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc3deca0-6cd1-43f6-84f7-98b600eaa4e7-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 10:22:38 crc kubenswrapper[4787]: I1001 10:22:38.562704 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc3deca0-6cd1-43f6-84f7-98b600eaa4e7-kube-api-access-2rnkn" (OuterVolumeSpecName: "kube-api-access-2rnkn") pod "bc3deca0-6cd1-43f6-84f7-98b600eaa4e7" (UID: "bc3deca0-6cd1-43f6-84f7-98b600eaa4e7"). InnerVolumeSpecName "kube-api-access-2rnkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:22:38 crc kubenswrapper[4787]: I1001 10:22:38.568512 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc3deca0-6cd1-43f6-84f7-98b600eaa4e7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bc3deca0-6cd1-43f6-84f7-98b600eaa4e7" (UID: "bc3deca0-6cd1-43f6-84f7-98b600eaa4e7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:22:38 crc kubenswrapper[4787]: I1001 10:22:38.657063 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc3deca0-6cd1-43f6-84f7-98b600eaa4e7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 10:22:38 crc kubenswrapper[4787]: I1001 10:22:38.657128 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2rnkn\" (UniqueName: \"kubernetes.io/projected/bc3deca0-6cd1-43f6-84f7-98b600eaa4e7-kube-api-access-2rnkn\") on node \"crc\" DevicePath \"\"" Oct 01 10:22:38 crc kubenswrapper[4787]: I1001 10:22:38.933954 4787 generic.go:334] "Generic (PLEG): container finished" podID="bc3deca0-6cd1-43f6-84f7-98b600eaa4e7" containerID="4387d1306a4a3ef35860c3b5e8dece9d951d3dfa943369c82d3242074abc3cea" exitCode=0 Oct 01 10:22:38 crc kubenswrapper[4787]: I1001 10:22:38.934004 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w4jbs" event={"ID":"bc3deca0-6cd1-43f6-84f7-98b600eaa4e7","Type":"ContainerDied","Data":"4387d1306a4a3ef35860c3b5e8dece9d951d3dfa943369c82d3242074abc3cea"} Oct 01 10:22:38 crc kubenswrapper[4787]: I1001 10:22:38.934033 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w4jbs" event={"ID":"bc3deca0-6cd1-43f6-84f7-98b600eaa4e7","Type":"ContainerDied","Data":"40cb4b049d239567ffe6a88066bd5e7dc4772a7b71421936bde81299b729f7d0"} Oct 01 10:22:38 crc kubenswrapper[4787]: I1001 10:22:38.934051 4787 scope.go:117] "RemoveContainer" containerID="4387d1306a4a3ef35860c3b5e8dece9d951d3dfa943369c82d3242074abc3cea" Oct 01 10:22:38 crc kubenswrapper[4787]: I1001 10:22:38.934187 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w4jbs" Oct 01 10:22:38 crc kubenswrapper[4787]: I1001 10:22:38.957720 4787 scope.go:117] "RemoveContainer" containerID="860b7cb0b20f12a2ac5b3022980865dc8f97f5da1930da8f6ab0c2c9cfd344d5" Oct 01 10:22:38 crc kubenswrapper[4787]: I1001 10:22:38.971876 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w4jbs"] Oct 01 10:22:38 crc kubenswrapper[4787]: I1001 10:22:38.978427 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-w4jbs"] Oct 01 10:22:38 crc kubenswrapper[4787]: I1001 10:22:38.994757 4787 scope.go:117] "RemoveContainer" containerID="919d5e817538ec492a53fe5c0c7bc7eeb5fe1c706fc140027eea45ee5fe106af" Oct 01 10:22:39 crc kubenswrapper[4787]: I1001 10:22:39.028574 4787 scope.go:117] "RemoveContainer" containerID="4387d1306a4a3ef35860c3b5e8dece9d951d3dfa943369c82d3242074abc3cea" Oct 01 10:22:39 crc kubenswrapper[4787]: E1001 10:22:39.028953 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4387d1306a4a3ef35860c3b5e8dece9d951d3dfa943369c82d3242074abc3cea\": container with ID starting with 4387d1306a4a3ef35860c3b5e8dece9d951d3dfa943369c82d3242074abc3cea not found: ID does not exist" containerID="4387d1306a4a3ef35860c3b5e8dece9d951d3dfa943369c82d3242074abc3cea" Oct 01 10:22:39 crc kubenswrapper[4787]: I1001 10:22:39.028987 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4387d1306a4a3ef35860c3b5e8dece9d951d3dfa943369c82d3242074abc3cea"} err="failed to get container status \"4387d1306a4a3ef35860c3b5e8dece9d951d3dfa943369c82d3242074abc3cea\": rpc error: code = NotFound desc = could not find container \"4387d1306a4a3ef35860c3b5e8dece9d951d3dfa943369c82d3242074abc3cea\": container with ID starting with 4387d1306a4a3ef35860c3b5e8dece9d951d3dfa943369c82d3242074abc3cea not found: ID does not exist" Oct 01 10:22:39 crc kubenswrapper[4787]: I1001 10:22:39.029008 4787 scope.go:117] "RemoveContainer" containerID="860b7cb0b20f12a2ac5b3022980865dc8f97f5da1930da8f6ab0c2c9cfd344d5" Oct 01 10:22:39 crc kubenswrapper[4787]: E1001 10:22:39.029494 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"860b7cb0b20f12a2ac5b3022980865dc8f97f5da1930da8f6ab0c2c9cfd344d5\": container with ID starting with 860b7cb0b20f12a2ac5b3022980865dc8f97f5da1930da8f6ab0c2c9cfd344d5 not found: ID does not exist" containerID="860b7cb0b20f12a2ac5b3022980865dc8f97f5da1930da8f6ab0c2c9cfd344d5" Oct 01 10:22:39 crc kubenswrapper[4787]: I1001 10:22:39.029521 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"860b7cb0b20f12a2ac5b3022980865dc8f97f5da1930da8f6ab0c2c9cfd344d5"} err="failed to get container status \"860b7cb0b20f12a2ac5b3022980865dc8f97f5da1930da8f6ab0c2c9cfd344d5\": rpc error: code = NotFound desc = could not find container \"860b7cb0b20f12a2ac5b3022980865dc8f97f5da1930da8f6ab0c2c9cfd344d5\": container with ID starting with 860b7cb0b20f12a2ac5b3022980865dc8f97f5da1930da8f6ab0c2c9cfd344d5 not found: ID does not exist" Oct 01 10:22:39 crc kubenswrapper[4787]: I1001 10:22:39.029538 4787 scope.go:117] "RemoveContainer" containerID="919d5e817538ec492a53fe5c0c7bc7eeb5fe1c706fc140027eea45ee5fe106af" Oct 01 10:22:39 crc kubenswrapper[4787]: E1001 10:22:39.029897 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"919d5e817538ec492a53fe5c0c7bc7eeb5fe1c706fc140027eea45ee5fe106af\": container with ID starting with 919d5e817538ec492a53fe5c0c7bc7eeb5fe1c706fc140027eea45ee5fe106af not found: ID does not exist" containerID="919d5e817538ec492a53fe5c0c7bc7eeb5fe1c706fc140027eea45ee5fe106af" Oct 01 10:22:39 crc kubenswrapper[4787]: I1001 10:22:39.029919 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"919d5e817538ec492a53fe5c0c7bc7eeb5fe1c706fc140027eea45ee5fe106af"} err="failed to get container status \"919d5e817538ec492a53fe5c0c7bc7eeb5fe1c706fc140027eea45ee5fe106af\": rpc error: code = NotFound desc = could not find container \"919d5e817538ec492a53fe5c0c7bc7eeb5fe1c706fc140027eea45ee5fe106af\": container with ID starting with 919d5e817538ec492a53fe5c0c7bc7eeb5fe1c706fc140027eea45ee5fe106af not found: ID does not exist" Oct 01 10:22:40 crc kubenswrapper[4787]: I1001 10:22:40.535771 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc3deca0-6cd1-43f6-84f7-98b600eaa4e7" path="/var/lib/kubelet/pods/bc3deca0-6cd1-43f6-84f7-98b600eaa4e7/volumes" Oct 01 10:22:41 crc kubenswrapper[4787]: I1001 10:22:41.250829 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:22:41 crc kubenswrapper[4787]: I1001 10:22:41.251220 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:23:11 crc kubenswrapper[4787]: I1001 10:23:11.251212 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:23:11 crc kubenswrapper[4787]: I1001 10:23:11.251853 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:23:41 crc kubenswrapper[4787]: I1001 10:23:41.250236 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:23:41 crc kubenswrapper[4787]: I1001 10:23:41.250758 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:23:41 crc kubenswrapper[4787]: I1001 10:23:41.250806 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" Oct 01 10:23:41 crc kubenswrapper[4787]: I1001 10:23:41.251671 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3df952eaa93a8322715e6a8d23a95eedf91662575215810014140a3e7d2a03bd"} pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 10:23:41 crc kubenswrapper[4787]: I1001 10:23:41.251722 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" containerID="cri-o://3df952eaa93a8322715e6a8d23a95eedf91662575215810014140a3e7d2a03bd" gracePeriod=600 Oct 01 10:23:41 crc kubenswrapper[4787]: E1001 10:23:41.421423 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:23:41 crc kubenswrapper[4787]: I1001 10:23:41.499811 4787 generic.go:334] "Generic (PLEG): container finished" podID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerID="3df952eaa93a8322715e6a8d23a95eedf91662575215810014140a3e7d2a03bd" exitCode=0 Oct 01 10:23:41 crc kubenswrapper[4787]: I1001 10:23:41.499887 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" event={"ID":"fa51b883-5f3e-4141-9d4c-37704aac7718","Type":"ContainerDied","Data":"3df952eaa93a8322715e6a8d23a95eedf91662575215810014140a3e7d2a03bd"} Oct 01 10:23:41 crc kubenswrapper[4787]: I1001 10:23:41.499972 4787 scope.go:117] "RemoveContainer" containerID="f6c4f2a4060e1ca1135f9288e823ff4cd4277a621cbd921d70bcb532fa4e420a" Oct 01 10:23:41 crc kubenswrapper[4787]: I1001 10:23:41.501383 4787 scope.go:117] "RemoveContainer" containerID="3df952eaa93a8322715e6a8d23a95eedf91662575215810014140a3e7d2a03bd" Oct 01 10:23:41 crc kubenswrapper[4787]: E1001 10:23:41.501974 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:23:53 crc kubenswrapper[4787]: I1001 10:23:53.524485 4787 scope.go:117] "RemoveContainer" containerID="3df952eaa93a8322715e6a8d23a95eedf91662575215810014140a3e7d2a03bd" Oct 01 10:23:53 crc kubenswrapper[4787]: E1001 10:23:53.525348 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:24:04 crc kubenswrapper[4787]: I1001 10:24:04.524284 4787 scope.go:117] "RemoveContainer" containerID="3df952eaa93a8322715e6a8d23a95eedf91662575215810014140a3e7d2a03bd" Oct 01 10:24:04 crc kubenswrapper[4787]: E1001 10:24:04.526468 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:24:19 crc kubenswrapper[4787]: I1001 10:24:19.523929 4787 scope.go:117] "RemoveContainer" containerID="3df952eaa93a8322715e6a8d23a95eedf91662575215810014140a3e7d2a03bd" Oct 01 10:24:19 crc kubenswrapper[4787]: E1001 10:24:19.524748 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:24:31 crc kubenswrapper[4787]: I1001 10:24:31.524354 4787 scope.go:117] "RemoveContainer" containerID="3df952eaa93a8322715e6a8d23a95eedf91662575215810014140a3e7d2a03bd" Oct 01 10:24:31 crc kubenswrapper[4787]: E1001 10:24:31.525433 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:24:44 crc kubenswrapper[4787]: I1001 10:24:44.525388 4787 scope.go:117] "RemoveContainer" containerID="3df952eaa93a8322715e6a8d23a95eedf91662575215810014140a3e7d2a03bd" Oct 01 10:24:44 crc kubenswrapper[4787]: E1001 10:24:44.527300 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:24:55 crc kubenswrapper[4787]: I1001 10:24:55.525152 4787 scope.go:117] "RemoveContainer" containerID="3df952eaa93a8322715e6a8d23a95eedf91662575215810014140a3e7d2a03bd" Oct 01 10:24:55 crc kubenswrapper[4787]: E1001 10:24:55.526712 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:25:07 crc kubenswrapper[4787]: I1001 10:25:07.524314 4787 scope.go:117] "RemoveContainer" containerID="3df952eaa93a8322715e6a8d23a95eedf91662575215810014140a3e7d2a03bd" Oct 01 10:25:07 crc kubenswrapper[4787]: E1001 10:25:07.525168 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:25:18 crc kubenswrapper[4787]: I1001 10:25:18.531532 4787 scope.go:117] "RemoveContainer" containerID="3df952eaa93a8322715e6a8d23a95eedf91662575215810014140a3e7d2a03bd" Oct 01 10:25:18 crc kubenswrapper[4787]: E1001 10:25:18.533019 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:25:30 crc kubenswrapper[4787]: I1001 10:25:30.523693 4787 scope.go:117] "RemoveContainer" containerID="3df952eaa93a8322715e6a8d23a95eedf91662575215810014140a3e7d2a03bd" Oct 01 10:25:30 crc kubenswrapper[4787]: E1001 10:25:30.524717 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:25:44 crc kubenswrapper[4787]: I1001 10:25:44.525098 4787 scope.go:117] "RemoveContainer" containerID="3df952eaa93a8322715e6a8d23a95eedf91662575215810014140a3e7d2a03bd" Oct 01 10:25:44 crc kubenswrapper[4787]: E1001 10:25:44.525865 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:25:55 crc kubenswrapper[4787]: I1001 10:25:55.523965 4787 scope.go:117] "RemoveContainer" containerID="3df952eaa93a8322715e6a8d23a95eedf91662575215810014140a3e7d2a03bd" Oct 01 10:25:55 crc kubenswrapper[4787]: E1001 10:25:55.525004 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:26:09 crc kubenswrapper[4787]: I1001 10:26:09.525113 4787 scope.go:117] "RemoveContainer" containerID="3df952eaa93a8322715e6a8d23a95eedf91662575215810014140a3e7d2a03bd" Oct 01 10:26:09 crc kubenswrapper[4787]: E1001 10:26:09.526191 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:26:24 crc kubenswrapper[4787]: I1001 10:26:24.525144 4787 scope.go:117] "RemoveContainer" containerID="3df952eaa93a8322715e6a8d23a95eedf91662575215810014140a3e7d2a03bd" Oct 01 10:26:24 crc kubenswrapper[4787]: E1001 10:26:24.526716 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:26:37 crc kubenswrapper[4787]: I1001 10:26:37.524518 4787 scope.go:117] "RemoveContainer" containerID="3df952eaa93a8322715e6a8d23a95eedf91662575215810014140a3e7d2a03bd" Oct 01 10:26:37 crc kubenswrapper[4787]: E1001 10:26:37.525448 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:26:49 crc kubenswrapper[4787]: I1001 10:26:49.524553 4787 scope.go:117] "RemoveContainer" containerID="3df952eaa93a8322715e6a8d23a95eedf91662575215810014140a3e7d2a03bd" Oct 01 10:26:49 crc kubenswrapper[4787]: E1001 10:26:49.526616 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:27:04 crc kubenswrapper[4787]: I1001 10:27:04.524930 4787 scope.go:117] "RemoveContainer" containerID="3df952eaa93a8322715e6a8d23a95eedf91662575215810014140a3e7d2a03bd" Oct 01 10:27:04 crc kubenswrapper[4787]: E1001 10:27:04.525767 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:27:18 crc kubenswrapper[4787]: I1001 10:27:18.531472 4787 scope.go:117] "RemoveContainer" containerID="3df952eaa93a8322715e6a8d23a95eedf91662575215810014140a3e7d2a03bd" Oct 01 10:27:18 crc kubenswrapper[4787]: E1001 10:27:18.533449 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:27:29 crc kubenswrapper[4787]: I1001 10:27:29.524323 4787 scope.go:117] "RemoveContainer" containerID="3df952eaa93a8322715e6a8d23a95eedf91662575215810014140a3e7d2a03bd" Oct 01 10:27:29 crc kubenswrapper[4787]: E1001 10:27:29.525651 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:27:42 crc kubenswrapper[4787]: I1001 10:27:42.524406 4787 scope.go:117] "RemoveContainer" containerID="3df952eaa93a8322715e6a8d23a95eedf91662575215810014140a3e7d2a03bd" Oct 01 10:27:42 crc kubenswrapper[4787]: E1001 10:27:42.525109 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:27:57 crc kubenswrapper[4787]: I1001 10:27:57.525244 4787 scope.go:117] "RemoveContainer" containerID="3df952eaa93a8322715e6a8d23a95eedf91662575215810014140a3e7d2a03bd" Oct 01 10:27:57 crc kubenswrapper[4787]: E1001 10:27:57.526852 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:28:12 crc kubenswrapper[4787]: I1001 10:28:12.524002 4787 scope.go:117] "RemoveContainer" containerID="3df952eaa93a8322715e6a8d23a95eedf91662575215810014140a3e7d2a03bd" Oct 01 10:28:12 crc kubenswrapper[4787]: E1001 10:28:12.525020 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:28:24 crc kubenswrapper[4787]: I1001 10:28:24.523663 4787 scope.go:117] "RemoveContainer" containerID="3df952eaa93a8322715e6a8d23a95eedf91662575215810014140a3e7d2a03bd" Oct 01 10:28:24 crc kubenswrapper[4787]: E1001 10:28:24.524585 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:28:38 crc kubenswrapper[4787]: I1001 10:28:38.531162 4787 scope.go:117] "RemoveContainer" containerID="3df952eaa93a8322715e6a8d23a95eedf91662575215810014140a3e7d2a03bd" Oct 01 10:28:38 crc kubenswrapper[4787]: E1001 10:28:38.532042 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:28:49 crc kubenswrapper[4787]: I1001 10:28:49.524119 4787 scope.go:117] "RemoveContainer" containerID="3df952eaa93a8322715e6a8d23a95eedf91662575215810014140a3e7d2a03bd" Oct 01 10:28:50 crc kubenswrapper[4787]: I1001 10:28:50.455737 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" event={"ID":"fa51b883-5f3e-4141-9d4c-37704aac7718","Type":"ContainerStarted","Data":"600ca70fa1dc39a1299d76f2c60704fbc1ec67b77ac9063a4f3e5239fa4265de"} Oct 01 10:30:00 crc kubenswrapper[4787]: I1001 10:30:00.211861 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321910-gfvnr"] Oct 01 10:30:00 crc kubenswrapper[4787]: E1001 10:30:00.218056 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc3deca0-6cd1-43f6-84f7-98b600eaa4e7" containerName="extract-utilities" Oct 01 10:30:00 crc kubenswrapper[4787]: I1001 10:30:00.218093 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc3deca0-6cd1-43f6-84f7-98b600eaa4e7" containerName="extract-utilities" Oct 01 10:30:00 crc kubenswrapper[4787]: E1001 10:30:00.218115 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc3deca0-6cd1-43f6-84f7-98b600eaa4e7" containerName="registry-server" Oct 01 10:30:00 crc kubenswrapper[4787]: I1001 10:30:00.218123 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc3deca0-6cd1-43f6-84f7-98b600eaa4e7" containerName="registry-server" Oct 01 10:30:00 crc kubenswrapper[4787]: E1001 10:30:00.218175 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc3deca0-6cd1-43f6-84f7-98b600eaa4e7" containerName="extract-content" Oct 01 10:30:00 crc kubenswrapper[4787]: I1001 10:30:00.218182 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc3deca0-6cd1-43f6-84f7-98b600eaa4e7" containerName="extract-content" Oct 01 10:30:00 crc kubenswrapper[4787]: I1001 10:30:00.218692 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc3deca0-6cd1-43f6-84f7-98b600eaa4e7" containerName="registry-server" Oct 01 10:30:00 crc kubenswrapper[4787]: I1001 10:30:00.219722 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321910-gfvnr" Oct 01 10:30:00 crc kubenswrapper[4787]: I1001 10:30:00.222706 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 01 10:30:00 crc kubenswrapper[4787]: I1001 10:30:00.224955 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 01 10:30:00 crc kubenswrapper[4787]: I1001 10:30:00.226930 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321910-gfvnr"] Oct 01 10:30:00 crc kubenswrapper[4787]: I1001 10:30:00.369150 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/08c32ec7-f42e-4d82-8e5e-d759f64b387c-secret-volume\") pod \"collect-profiles-29321910-gfvnr\" (UID: \"08c32ec7-f42e-4d82-8e5e-d759f64b387c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321910-gfvnr" Oct 01 10:30:00 crc kubenswrapper[4787]: I1001 10:30:00.369306 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwkbx\" (UniqueName: \"kubernetes.io/projected/08c32ec7-f42e-4d82-8e5e-d759f64b387c-kube-api-access-jwkbx\") pod \"collect-profiles-29321910-gfvnr\" (UID: \"08c32ec7-f42e-4d82-8e5e-d759f64b387c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321910-gfvnr" Oct 01 10:30:00 crc kubenswrapper[4787]: I1001 10:30:00.369329 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/08c32ec7-f42e-4d82-8e5e-d759f64b387c-config-volume\") pod \"collect-profiles-29321910-gfvnr\" (UID: \"08c32ec7-f42e-4d82-8e5e-d759f64b387c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321910-gfvnr" Oct 01 10:30:00 crc kubenswrapper[4787]: I1001 10:30:00.471761 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwkbx\" (UniqueName: \"kubernetes.io/projected/08c32ec7-f42e-4d82-8e5e-d759f64b387c-kube-api-access-jwkbx\") pod \"collect-profiles-29321910-gfvnr\" (UID: \"08c32ec7-f42e-4d82-8e5e-d759f64b387c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321910-gfvnr" Oct 01 10:30:00 crc kubenswrapper[4787]: I1001 10:30:00.472104 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/08c32ec7-f42e-4d82-8e5e-d759f64b387c-config-volume\") pod \"collect-profiles-29321910-gfvnr\" (UID: \"08c32ec7-f42e-4d82-8e5e-d759f64b387c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321910-gfvnr" Oct 01 10:30:00 crc kubenswrapper[4787]: I1001 10:30:00.472326 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/08c32ec7-f42e-4d82-8e5e-d759f64b387c-secret-volume\") pod \"collect-profiles-29321910-gfvnr\" (UID: \"08c32ec7-f42e-4d82-8e5e-d759f64b387c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321910-gfvnr" Oct 01 10:30:00 crc kubenswrapper[4787]: I1001 10:30:00.472945 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/08c32ec7-f42e-4d82-8e5e-d759f64b387c-config-volume\") pod \"collect-profiles-29321910-gfvnr\" (UID: \"08c32ec7-f42e-4d82-8e5e-d759f64b387c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321910-gfvnr" Oct 01 10:30:00 crc kubenswrapper[4787]: I1001 10:30:00.479865 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/08c32ec7-f42e-4d82-8e5e-d759f64b387c-secret-volume\") pod \"collect-profiles-29321910-gfvnr\" (UID: \"08c32ec7-f42e-4d82-8e5e-d759f64b387c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321910-gfvnr" Oct 01 10:30:00 crc kubenswrapper[4787]: I1001 10:30:00.490667 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwkbx\" (UniqueName: \"kubernetes.io/projected/08c32ec7-f42e-4d82-8e5e-d759f64b387c-kube-api-access-jwkbx\") pod \"collect-profiles-29321910-gfvnr\" (UID: \"08c32ec7-f42e-4d82-8e5e-d759f64b387c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321910-gfvnr" Oct 01 10:30:00 crc kubenswrapper[4787]: I1001 10:30:00.556916 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321910-gfvnr" Oct 01 10:30:01 crc kubenswrapper[4787]: I1001 10:30:01.008318 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321910-gfvnr"] Oct 01 10:30:01 crc kubenswrapper[4787]: I1001 10:30:01.100769 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321910-gfvnr" event={"ID":"08c32ec7-f42e-4d82-8e5e-d759f64b387c","Type":"ContainerStarted","Data":"06dba139034c184855db85aa4a21398949e88f1ecbbaec63bec4e5c014c126d4"} Oct 01 10:30:02 crc kubenswrapper[4787]: I1001 10:30:02.119741 4787 generic.go:334] "Generic (PLEG): container finished" podID="08c32ec7-f42e-4d82-8e5e-d759f64b387c" containerID="fa800703ed4ee6a9c870c4b78c4660c159e7d60078085a2ee06221495ccce6d0" exitCode=0 Oct 01 10:30:02 crc kubenswrapper[4787]: I1001 10:30:02.119844 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321910-gfvnr" event={"ID":"08c32ec7-f42e-4d82-8e5e-d759f64b387c","Type":"ContainerDied","Data":"fa800703ed4ee6a9c870c4b78c4660c159e7d60078085a2ee06221495ccce6d0"} Oct 01 10:30:03 crc kubenswrapper[4787]: I1001 10:30:03.491584 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321910-gfvnr" Oct 01 10:30:03 crc kubenswrapper[4787]: I1001 10:30:03.641221 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/08c32ec7-f42e-4d82-8e5e-d759f64b387c-config-volume\") pod \"08c32ec7-f42e-4d82-8e5e-d759f64b387c\" (UID: \"08c32ec7-f42e-4d82-8e5e-d759f64b387c\") " Oct 01 10:30:03 crc kubenswrapper[4787]: I1001 10:30:03.641474 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/08c32ec7-f42e-4d82-8e5e-d759f64b387c-secret-volume\") pod \"08c32ec7-f42e-4d82-8e5e-d759f64b387c\" (UID: \"08c32ec7-f42e-4d82-8e5e-d759f64b387c\") " Oct 01 10:30:03 crc kubenswrapper[4787]: I1001 10:30:03.641606 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jwkbx\" (UniqueName: \"kubernetes.io/projected/08c32ec7-f42e-4d82-8e5e-d759f64b387c-kube-api-access-jwkbx\") pod \"08c32ec7-f42e-4d82-8e5e-d759f64b387c\" (UID: \"08c32ec7-f42e-4d82-8e5e-d759f64b387c\") " Oct 01 10:30:03 crc kubenswrapper[4787]: I1001 10:30:03.641824 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08c32ec7-f42e-4d82-8e5e-d759f64b387c-config-volume" (OuterVolumeSpecName: "config-volume") pod "08c32ec7-f42e-4d82-8e5e-d759f64b387c" (UID: "08c32ec7-f42e-4d82-8e5e-d759f64b387c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:30:03 crc kubenswrapper[4787]: I1001 10:30:03.642682 4787 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/08c32ec7-f42e-4d82-8e5e-d759f64b387c-config-volume\") on node \"crc\" DevicePath \"\"" Oct 01 10:30:03 crc kubenswrapper[4787]: I1001 10:30:03.651437 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08c32ec7-f42e-4d82-8e5e-d759f64b387c-kube-api-access-jwkbx" (OuterVolumeSpecName: "kube-api-access-jwkbx") pod "08c32ec7-f42e-4d82-8e5e-d759f64b387c" (UID: "08c32ec7-f42e-4d82-8e5e-d759f64b387c"). InnerVolumeSpecName "kube-api-access-jwkbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:30:03 crc kubenswrapper[4787]: I1001 10:30:03.651450 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08c32ec7-f42e-4d82-8e5e-d759f64b387c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "08c32ec7-f42e-4d82-8e5e-d759f64b387c" (UID: "08c32ec7-f42e-4d82-8e5e-d759f64b387c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:30:03 crc kubenswrapper[4787]: I1001 10:30:03.744925 4787 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/08c32ec7-f42e-4d82-8e5e-d759f64b387c-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 01 10:30:03 crc kubenswrapper[4787]: I1001 10:30:03.744966 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jwkbx\" (UniqueName: \"kubernetes.io/projected/08c32ec7-f42e-4d82-8e5e-d759f64b387c-kube-api-access-jwkbx\") on node \"crc\" DevicePath \"\"" Oct 01 10:30:04 crc kubenswrapper[4787]: I1001 10:30:04.140480 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321910-gfvnr" event={"ID":"08c32ec7-f42e-4d82-8e5e-d759f64b387c","Type":"ContainerDied","Data":"06dba139034c184855db85aa4a21398949e88f1ecbbaec63bec4e5c014c126d4"} Oct 01 10:30:04 crc kubenswrapper[4787]: I1001 10:30:04.141019 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="06dba139034c184855db85aa4a21398949e88f1ecbbaec63bec4e5c014c126d4" Oct 01 10:30:04 crc kubenswrapper[4787]: I1001 10:30:04.140577 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321910-gfvnr" Oct 01 10:30:04 crc kubenswrapper[4787]: I1001 10:30:04.565878 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321865-9nl5h"] Oct 01 10:30:04 crc kubenswrapper[4787]: I1001 10:30:04.579760 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321865-9nl5h"] Oct 01 10:30:06 crc kubenswrapper[4787]: I1001 10:30:06.534922 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c159d29b-be19-4d0f-980e-a12c2ed641f2" path="/var/lib/kubelet/pods/c159d29b-be19-4d0f-980e-a12c2ed641f2/volumes" Oct 01 10:30:33 crc kubenswrapper[4787]: I1001 10:30:33.085870 4787 scope.go:117] "RemoveContainer" containerID="863b0a8fe30dcfd7f3f0f2313758f642a66355859da1809cf9417915270239f0" Oct 01 10:31:11 crc kubenswrapper[4787]: I1001 10:31:11.250822 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:31:11 crc kubenswrapper[4787]: I1001 10:31:11.251488 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:31:12 crc kubenswrapper[4787]: I1001 10:31:12.805817 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hf9fh"] Oct 01 10:31:12 crc kubenswrapper[4787]: E1001 10:31:12.806252 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08c32ec7-f42e-4d82-8e5e-d759f64b387c" containerName="collect-profiles" Oct 01 10:31:12 crc kubenswrapper[4787]: I1001 10:31:12.806265 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="08c32ec7-f42e-4d82-8e5e-d759f64b387c" containerName="collect-profiles" Oct 01 10:31:12 crc kubenswrapper[4787]: I1001 10:31:12.806467 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="08c32ec7-f42e-4d82-8e5e-d759f64b387c" containerName="collect-profiles" Oct 01 10:31:12 crc kubenswrapper[4787]: I1001 10:31:12.810919 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hf9fh" Oct 01 10:31:12 crc kubenswrapper[4787]: I1001 10:31:12.836000 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hf9fh"] Oct 01 10:31:12 crc kubenswrapper[4787]: I1001 10:31:12.989469 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2xm9\" (UniqueName: \"kubernetes.io/projected/d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68-kube-api-access-m2xm9\") pod \"community-operators-hf9fh\" (UID: \"d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68\") " pod="openshift-marketplace/community-operators-hf9fh" Oct 01 10:31:12 crc kubenswrapper[4787]: I1001 10:31:12.989960 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68-catalog-content\") pod \"community-operators-hf9fh\" (UID: \"d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68\") " pod="openshift-marketplace/community-operators-hf9fh" Oct 01 10:31:12 crc kubenswrapper[4787]: I1001 10:31:12.990126 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68-utilities\") pod \"community-operators-hf9fh\" (UID: \"d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68\") " pod="openshift-marketplace/community-operators-hf9fh" Oct 01 10:31:13 crc kubenswrapper[4787]: I1001 10:31:13.091844 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68-catalog-content\") pod \"community-operators-hf9fh\" (UID: \"d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68\") " pod="openshift-marketplace/community-operators-hf9fh" Oct 01 10:31:13 crc kubenswrapper[4787]: I1001 10:31:13.091950 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68-utilities\") pod \"community-operators-hf9fh\" (UID: \"d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68\") " pod="openshift-marketplace/community-operators-hf9fh" Oct 01 10:31:13 crc kubenswrapper[4787]: I1001 10:31:13.092007 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2xm9\" (UniqueName: \"kubernetes.io/projected/d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68-kube-api-access-m2xm9\") pod \"community-operators-hf9fh\" (UID: \"d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68\") " pod="openshift-marketplace/community-operators-hf9fh" Oct 01 10:31:13 crc kubenswrapper[4787]: I1001 10:31:13.092450 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68-catalog-content\") pod \"community-operators-hf9fh\" (UID: \"d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68\") " pod="openshift-marketplace/community-operators-hf9fh" Oct 01 10:31:13 crc kubenswrapper[4787]: I1001 10:31:13.092680 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68-utilities\") pod \"community-operators-hf9fh\" (UID: \"d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68\") " pod="openshift-marketplace/community-operators-hf9fh" Oct 01 10:31:13 crc kubenswrapper[4787]: I1001 10:31:13.118968 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2xm9\" (UniqueName: \"kubernetes.io/projected/d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68-kube-api-access-m2xm9\") pod \"community-operators-hf9fh\" (UID: \"d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68\") " pod="openshift-marketplace/community-operators-hf9fh" Oct 01 10:31:13 crc kubenswrapper[4787]: I1001 10:31:13.160891 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hf9fh" Oct 01 10:31:13 crc kubenswrapper[4787]: I1001 10:31:13.716918 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hf9fh"] Oct 01 10:31:13 crc kubenswrapper[4787]: I1001 10:31:13.791661 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hf9fh" event={"ID":"d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68","Type":"ContainerStarted","Data":"5d765e0c6de773cd70d76add89126de68efb1ebe432c0dc5e99a2d94c28dca39"} Oct 01 10:31:13 crc kubenswrapper[4787]: I1001 10:31:13.794596 4787 generic.go:334] "Generic (PLEG): container finished" podID="4b136104-8797-40d5-9ea1-fec8c6be0e90" containerID="1221312c3fea241a9c775a3b4f5142913e44ed581fc20935359dbbd60a68af15" exitCode=0 Oct 01 10:31:13 crc kubenswrapper[4787]: I1001 10:31:13.794672 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"4b136104-8797-40d5-9ea1-fec8c6be0e90","Type":"ContainerDied","Data":"1221312c3fea241a9c775a3b4f5142913e44ed581fc20935359dbbd60a68af15"} Oct 01 10:31:14 crc kubenswrapper[4787]: I1001 10:31:14.809619 4787 generic.go:334] "Generic (PLEG): container finished" podID="d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68" containerID="c49e34c37f7c12cc823a38af312499737b17949a55533f0a3b1d84287a2e2c2f" exitCode=0 Oct 01 10:31:14 crc kubenswrapper[4787]: I1001 10:31:14.809725 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hf9fh" event={"ID":"d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68","Type":"ContainerDied","Data":"c49e34c37f7c12cc823a38af312499737b17949a55533f0a3b1d84287a2e2c2f"} Oct 01 10:31:14 crc kubenswrapper[4787]: I1001 10:31:14.811986 4787 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 10:31:15 crc kubenswrapper[4787]: I1001 10:31:15.152447 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 01 10:31:15 crc kubenswrapper[4787]: I1001 10:31:15.335421 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4b136104-8797-40d5-9ea1-fec8c6be0e90-config-data\") pod \"4b136104-8797-40d5-9ea1-fec8c6be0e90\" (UID: \"4b136104-8797-40d5-9ea1-fec8c6be0e90\") " Oct 01 10:31:15 crc kubenswrapper[4787]: I1001 10:31:15.335464 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"4b136104-8797-40d5-9ea1-fec8c6be0e90\" (UID: \"4b136104-8797-40d5-9ea1-fec8c6be0e90\") " Oct 01 10:31:15 crc kubenswrapper[4787]: I1001 10:31:15.335515 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4b136104-8797-40d5-9ea1-fec8c6be0e90-openstack-config\") pod \"4b136104-8797-40d5-9ea1-fec8c6be0e90\" (UID: \"4b136104-8797-40d5-9ea1-fec8c6be0e90\") " Oct 01 10:31:15 crc kubenswrapper[4787]: I1001 10:31:15.335617 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4b136104-8797-40d5-9ea1-fec8c6be0e90-openstack-config-secret\") pod \"4b136104-8797-40d5-9ea1-fec8c6be0e90\" (UID: \"4b136104-8797-40d5-9ea1-fec8c6be0e90\") " Oct 01 10:31:15 crc kubenswrapper[4787]: I1001 10:31:15.335679 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/4b136104-8797-40d5-9ea1-fec8c6be0e90-test-operator-ephemeral-workdir\") pod \"4b136104-8797-40d5-9ea1-fec8c6be0e90\" (UID: \"4b136104-8797-40d5-9ea1-fec8c6be0e90\") " Oct 01 10:31:15 crc kubenswrapper[4787]: I1001 10:31:15.335747 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b136104-8797-40d5-9ea1-fec8c6be0e90-ssh-key\") pod \"4b136104-8797-40d5-9ea1-fec8c6be0e90\" (UID: \"4b136104-8797-40d5-9ea1-fec8c6be0e90\") " Oct 01 10:31:15 crc kubenswrapper[4787]: I1001 10:31:15.335808 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/4b136104-8797-40d5-9ea1-fec8c6be0e90-ca-certs\") pod \"4b136104-8797-40d5-9ea1-fec8c6be0e90\" (UID: \"4b136104-8797-40d5-9ea1-fec8c6be0e90\") " Oct 01 10:31:15 crc kubenswrapper[4787]: I1001 10:31:15.335864 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/4b136104-8797-40d5-9ea1-fec8c6be0e90-test-operator-ephemeral-temporary\") pod \"4b136104-8797-40d5-9ea1-fec8c6be0e90\" (UID: \"4b136104-8797-40d5-9ea1-fec8c6be0e90\") " Oct 01 10:31:15 crc kubenswrapper[4787]: I1001 10:31:15.335911 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gllb6\" (UniqueName: \"kubernetes.io/projected/4b136104-8797-40d5-9ea1-fec8c6be0e90-kube-api-access-gllb6\") pod \"4b136104-8797-40d5-9ea1-fec8c6be0e90\" (UID: \"4b136104-8797-40d5-9ea1-fec8c6be0e90\") " Oct 01 10:31:15 crc kubenswrapper[4787]: I1001 10:31:15.336285 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b136104-8797-40d5-9ea1-fec8c6be0e90-config-data" (OuterVolumeSpecName: "config-data") pod "4b136104-8797-40d5-9ea1-fec8c6be0e90" (UID: "4b136104-8797-40d5-9ea1-fec8c6be0e90"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:31:15 crc kubenswrapper[4787]: I1001 10:31:15.336419 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4b136104-8797-40d5-9ea1-fec8c6be0e90-config-data\") on node \"crc\" DevicePath \"\"" Oct 01 10:31:15 crc kubenswrapper[4787]: I1001 10:31:15.336712 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b136104-8797-40d5-9ea1-fec8c6be0e90-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "4b136104-8797-40d5-9ea1-fec8c6be0e90" (UID: "4b136104-8797-40d5-9ea1-fec8c6be0e90"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:31:15 crc kubenswrapper[4787]: I1001 10:31:15.340894 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b136104-8797-40d5-9ea1-fec8c6be0e90-kube-api-access-gllb6" (OuterVolumeSpecName: "kube-api-access-gllb6") pod "4b136104-8797-40d5-9ea1-fec8c6be0e90" (UID: "4b136104-8797-40d5-9ea1-fec8c6be0e90"). InnerVolumeSpecName "kube-api-access-gllb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:31:15 crc kubenswrapper[4787]: I1001 10:31:15.341706 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b136104-8797-40d5-9ea1-fec8c6be0e90-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "4b136104-8797-40d5-9ea1-fec8c6be0e90" (UID: "4b136104-8797-40d5-9ea1-fec8c6be0e90"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:31:15 crc kubenswrapper[4787]: I1001 10:31:15.341864 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "test-operator-logs") pod "4b136104-8797-40d5-9ea1-fec8c6be0e90" (UID: "4b136104-8797-40d5-9ea1-fec8c6be0e90"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 01 10:31:15 crc kubenswrapper[4787]: I1001 10:31:15.367446 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b136104-8797-40d5-9ea1-fec8c6be0e90-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "4b136104-8797-40d5-9ea1-fec8c6be0e90" (UID: "4b136104-8797-40d5-9ea1-fec8c6be0e90"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:31:15 crc kubenswrapper[4787]: I1001 10:31:15.367555 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b136104-8797-40d5-9ea1-fec8c6be0e90-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "4b136104-8797-40d5-9ea1-fec8c6be0e90" (UID: "4b136104-8797-40d5-9ea1-fec8c6be0e90"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:31:15 crc kubenswrapper[4787]: I1001 10:31:15.379633 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b136104-8797-40d5-9ea1-fec8c6be0e90-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4b136104-8797-40d5-9ea1-fec8c6be0e90" (UID: "4b136104-8797-40d5-9ea1-fec8c6be0e90"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:31:15 crc kubenswrapper[4787]: I1001 10:31:15.388220 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b136104-8797-40d5-9ea1-fec8c6be0e90-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "4b136104-8797-40d5-9ea1-fec8c6be0e90" (UID: "4b136104-8797-40d5-9ea1-fec8c6be0e90"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:31:15 crc kubenswrapper[4787]: I1001 10:31:15.438797 4787 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4b136104-8797-40d5-9ea1-fec8c6be0e90-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 01 10:31:15 crc kubenswrapper[4787]: I1001 10:31:15.438921 4787 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4b136104-8797-40d5-9ea1-fec8c6be0e90-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 01 10:31:15 crc kubenswrapper[4787]: I1001 10:31:15.438944 4787 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/4b136104-8797-40d5-9ea1-fec8c6be0e90-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 01 10:31:15 crc kubenswrapper[4787]: I1001 10:31:15.438957 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b136104-8797-40d5-9ea1-fec8c6be0e90-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 01 10:31:15 crc kubenswrapper[4787]: I1001 10:31:15.438970 4787 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/4b136104-8797-40d5-9ea1-fec8c6be0e90-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 01 10:31:15 crc kubenswrapper[4787]: I1001 10:31:15.438988 4787 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/4b136104-8797-40d5-9ea1-fec8c6be0e90-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 01 10:31:15 crc kubenswrapper[4787]: I1001 10:31:15.439004 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gllb6\" (UniqueName: \"kubernetes.io/projected/4b136104-8797-40d5-9ea1-fec8c6be0e90-kube-api-access-gllb6\") on node \"crc\" DevicePath \"\"" Oct 01 10:31:15 crc kubenswrapper[4787]: I1001 10:31:15.439052 4787 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Oct 01 10:31:15 crc kubenswrapper[4787]: I1001 10:31:15.468328 4787 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Oct 01 10:31:15 crc kubenswrapper[4787]: I1001 10:31:15.540099 4787 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Oct 01 10:31:15 crc kubenswrapper[4787]: I1001 10:31:15.824214 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"4b136104-8797-40d5-9ea1-fec8c6be0e90","Type":"ContainerDied","Data":"6042ac4649ddff4357aa290f8de50125f01a705ec06864e6086fe9777317fa00"} Oct 01 10:31:15 crc kubenswrapper[4787]: I1001 10:31:15.824523 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6042ac4649ddff4357aa290f8de50125f01a705ec06864e6086fe9777317fa00" Oct 01 10:31:15 crc kubenswrapper[4787]: I1001 10:31:15.824313 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 01 10:31:16 crc kubenswrapper[4787]: I1001 10:31:16.836879 4787 generic.go:334] "Generic (PLEG): container finished" podID="d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68" containerID="cf50eb1ed3aebedbb2446c620da051fa41a46eb823a79b05296dde3d002fddb7" exitCode=0 Oct 01 10:31:16 crc kubenswrapper[4787]: I1001 10:31:16.836944 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hf9fh" event={"ID":"d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68","Type":"ContainerDied","Data":"cf50eb1ed3aebedbb2446c620da051fa41a46eb823a79b05296dde3d002fddb7"} Oct 01 10:31:17 crc kubenswrapper[4787]: I1001 10:31:17.853749 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hf9fh" event={"ID":"d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68","Type":"ContainerStarted","Data":"e93c311ee809cafb338bd337ce971df314cc16e0f0563f30c209c6e87281eb43"} Oct 01 10:31:17 crc kubenswrapper[4787]: I1001 10:31:17.873022 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hf9fh" podStartSLOduration=3.4483452310000002 podStartE2EDuration="5.873001361s" podCreationTimestamp="2025-10-01 10:31:12 +0000 UTC" firstStartedPulling="2025-10-01 10:31:14.811674235 +0000 UTC m=+3306.926818392" lastFinishedPulling="2025-10-01 10:31:17.236330365 +0000 UTC m=+3309.351474522" observedRunningTime="2025-10-01 10:31:17.872649832 +0000 UTC m=+3309.987794019" watchObservedRunningTime="2025-10-01 10:31:17.873001361 +0000 UTC m=+3309.988145518" Oct 01 10:31:22 crc kubenswrapper[4787]: I1001 10:31:22.368634 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 01 10:31:22 crc kubenswrapper[4787]: E1001 10:31:22.370633 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b136104-8797-40d5-9ea1-fec8c6be0e90" containerName="tempest-tests-tempest-tests-runner" Oct 01 10:31:22 crc kubenswrapper[4787]: I1001 10:31:22.370701 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b136104-8797-40d5-9ea1-fec8c6be0e90" containerName="tempest-tests-tempest-tests-runner" Oct 01 10:31:22 crc kubenswrapper[4787]: I1001 10:31:22.371519 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b136104-8797-40d5-9ea1-fec8c6be0e90" containerName="tempest-tests-tempest-tests-runner" Oct 01 10:31:22 crc kubenswrapper[4787]: I1001 10:31:22.372999 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 01 10:31:22 crc kubenswrapper[4787]: I1001 10:31:22.377511 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-qq94k" Oct 01 10:31:22 crc kubenswrapper[4787]: I1001 10:31:22.395438 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 01 10:31:22 crc kubenswrapper[4787]: I1001 10:31:22.479990 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"9aaa5922-539c-4e98-99ac-a08a31b14d53\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 01 10:31:22 crc kubenswrapper[4787]: I1001 10:31:22.480102 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7crz\" (UniqueName: \"kubernetes.io/projected/9aaa5922-539c-4e98-99ac-a08a31b14d53-kube-api-access-p7crz\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"9aaa5922-539c-4e98-99ac-a08a31b14d53\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 01 10:31:22 crc kubenswrapper[4787]: I1001 10:31:22.581843 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"9aaa5922-539c-4e98-99ac-a08a31b14d53\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 01 10:31:22 crc kubenswrapper[4787]: I1001 10:31:22.581929 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7crz\" (UniqueName: \"kubernetes.io/projected/9aaa5922-539c-4e98-99ac-a08a31b14d53-kube-api-access-p7crz\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"9aaa5922-539c-4e98-99ac-a08a31b14d53\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 01 10:31:22 crc kubenswrapper[4787]: I1001 10:31:22.582359 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"9aaa5922-539c-4e98-99ac-a08a31b14d53\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 01 10:31:22 crc kubenswrapper[4787]: I1001 10:31:22.600701 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7crz\" (UniqueName: \"kubernetes.io/projected/9aaa5922-539c-4e98-99ac-a08a31b14d53-kube-api-access-p7crz\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"9aaa5922-539c-4e98-99ac-a08a31b14d53\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 01 10:31:22 crc kubenswrapper[4787]: I1001 10:31:22.613110 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"9aaa5922-539c-4e98-99ac-a08a31b14d53\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 01 10:31:22 crc kubenswrapper[4787]: I1001 10:31:22.705558 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 01 10:31:23 crc kubenswrapper[4787]: I1001 10:31:23.130666 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 01 10:31:23 crc kubenswrapper[4787]: W1001 10:31:23.139835 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9aaa5922_539c_4e98_99ac_a08a31b14d53.slice/crio-2317ace7f53d7caec19ae4aa280a0bdec8237bc3b71fd852446211a399498361 WatchSource:0}: Error finding container 2317ace7f53d7caec19ae4aa280a0bdec8237bc3b71fd852446211a399498361: Status 404 returned error can't find the container with id 2317ace7f53d7caec19ae4aa280a0bdec8237bc3b71fd852446211a399498361 Oct 01 10:31:23 crc kubenswrapper[4787]: I1001 10:31:23.162214 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hf9fh" Oct 01 10:31:23 crc kubenswrapper[4787]: I1001 10:31:23.162277 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hf9fh" Oct 01 10:31:23 crc kubenswrapper[4787]: I1001 10:31:23.208882 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hf9fh" Oct 01 10:31:23 crc kubenswrapper[4787]: I1001 10:31:23.906931 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"9aaa5922-539c-4e98-99ac-a08a31b14d53","Type":"ContainerStarted","Data":"2317ace7f53d7caec19ae4aa280a0bdec8237bc3b71fd852446211a399498361"} Oct 01 10:31:23 crc kubenswrapper[4787]: I1001 10:31:23.962441 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hf9fh" Oct 01 10:31:24 crc kubenswrapper[4787]: I1001 10:31:24.023421 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hf9fh"] Oct 01 10:31:25 crc kubenswrapper[4787]: I1001 10:31:25.926120 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hf9fh" podUID="d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68" containerName="registry-server" containerID="cri-o://e93c311ee809cafb338bd337ce971df314cc16e0f0563f30c209c6e87281eb43" gracePeriod=2 Oct 01 10:31:25 crc kubenswrapper[4787]: I1001 10:31:25.926599 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"9aaa5922-539c-4e98-99ac-a08a31b14d53","Type":"ContainerStarted","Data":"0f060217a15d60df9d22e8039c3050f226b3c9cf7d5c8aebfccff5ad982e49b9"} Oct 01 10:31:25 crc kubenswrapper[4787]: I1001 10:31:25.950343 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.890214745 podStartE2EDuration="3.950317568s" podCreationTimestamp="2025-10-01 10:31:22 +0000 UTC" firstStartedPulling="2025-10-01 10:31:23.142215162 +0000 UTC m=+3315.257359319" lastFinishedPulling="2025-10-01 10:31:25.202317985 +0000 UTC m=+3317.317462142" observedRunningTime="2025-10-01 10:31:25.941547323 +0000 UTC m=+3318.056691490" watchObservedRunningTime="2025-10-01 10:31:25.950317568 +0000 UTC m=+3318.065461735" Oct 01 10:31:26 crc kubenswrapper[4787]: I1001 10:31:26.936049 4787 generic.go:334] "Generic (PLEG): container finished" podID="d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68" containerID="e93c311ee809cafb338bd337ce971df314cc16e0f0563f30c209c6e87281eb43" exitCode=0 Oct 01 10:31:26 crc kubenswrapper[4787]: I1001 10:31:26.936123 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hf9fh" event={"ID":"d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68","Type":"ContainerDied","Data":"e93c311ee809cafb338bd337ce971df314cc16e0f0563f30c209c6e87281eb43"} Oct 01 10:31:27 crc kubenswrapper[4787]: I1001 10:31:27.063780 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hf9fh" Oct 01 10:31:27 crc kubenswrapper[4787]: I1001 10:31:27.170461 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2xm9\" (UniqueName: \"kubernetes.io/projected/d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68-kube-api-access-m2xm9\") pod \"d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68\" (UID: \"d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68\") " Oct 01 10:31:27 crc kubenswrapper[4787]: I1001 10:31:27.170635 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68-utilities\") pod \"d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68\" (UID: \"d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68\") " Oct 01 10:31:27 crc kubenswrapper[4787]: I1001 10:31:27.170737 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68-catalog-content\") pod \"d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68\" (UID: \"d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68\") " Oct 01 10:31:27 crc kubenswrapper[4787]: I1001 10:31:27.171694 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68-utilities" (OuterVolumeSpecName: "utilities") pod "d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68" (UID: "d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:31:27 crc kubenswrapper[4787]: I1001 10:31:27.176181 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68-kube-api-access-m2xm9" (OuterVolumeSpecName: "kube-api-access-m2xm9") pod "d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68" (UID: "d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68"). InnerVolumeSpecName "kube-api-access-m2xm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:31:27 crc kubenswrapper[4787]: I1001 10:31:27.273645 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 10:31:27 crc kubenswrapper[4787]: I1001 10:31:27.273707 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2xm9\" (UniqueName: \"kubernetes.io/projected/d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68-kube-api-access-m2xm9\") on node \"crc\" DevicePath \"\"" Oct 01 10:31:27 crc kubenswrapper[4787]: I1001 10:31:27.591186 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68" (UID: "d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:31:27 crc kubenswrapper[4787]: I1001 10:31:27.682648 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 10:31:27 crc kubenswrapper[4787]: I1001 10:31:27.947852 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hf9fh" event={"ID":"d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68","Type":"ContainerDied","Data":"5d765e0c6de773cd70d76add89126de68efb1ebe432c0dc5e99a2d94c28dca39"} Oct 01 10:31:27 crc kubenswrapper[4787]: I1001 10:31:27.947909 4787 scope.go:117] "RemoveContainer" containerID="e93c311ee809cafb338bd337ce971df314cc16e0f0563f30c209c6e87281eb43" Oct 01 10:31:27 crc kubenswrapper[4787]: I1001 10:31:27.947937 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hf9fh" Oct 01 10:31:27 crc kubenswrapper[4787]: I1001 10:31:27.970406 4787 scope.go:117] "RemoveContainer" containerID="cf50eb1ed3aebedbb2446c620da051fa41a46eb823a79b05296dde3d002fddb7" Oct 01 10:31:27 crc kubenswrapper[4787]: I1001 10:31:27.983315 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hf9fh"] Oct 01 10:31:27 crc kubenswrapper[4787]: I1001 10:31:27.990770 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hf9fh"] Oct 01 10:31:28 crc kubenswrapper[4787]: I1001 10:31:28.014502 4787 scope.go:117] "RemoveContainer" containerID="c49e34c37f7c12cc823a38af312499737b17949a55533f0a3b1d84287a2e2c2f" Oct 01 10:31:28 crc kubenswrapper[4787]: I1001 10:31:28.534489 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68" path="/var/lib/kubelet/pods/d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68/volumes" Oct 01 10:31:41 crc kubenswrapper[4787]: I1001 10:31:41.251391 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:31:41 crc kubenswrapper[4787]: I1001 10:31:41.252153 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:31:41 crc kubenswrapper[4787]: I1001 10:31:41.848306 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-4twfr/must-gather-89b86"] Oct 01 10:31:41 crc kubenswrapper[4787]: E1001 10:31:41.848985 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68" containerName="extract-content" Oct 01 10:31:41 crc kubenswrapper[4787]: I1001 10:31:41.849004 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68" containerName="extract-content" Oct 01 10:31:41 crc kubenswrapper[4787]: E1001 10:31:41.849018 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68" containerName="extract-utilities" Oct 01 10:31:41 crc kubenswrapper[4787]: I1001 10:31:41.849024 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68" containerName="extract-utilities" Oct 01 10:31:41 crc kubenswrapper[4787]: E1001 10:31:41.849054 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68" containerName="registry-server" Oct 01 10:31:41 crc kubenswrapper[4787]: I1001 10:31:41.849060 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68" containerName="registry-server" Oct 01 10:31:41 crc kubenswrapper[4787]: I1001 10:31:41.849895 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5b153cf-c59c-4f1d-8ca4-e575c2ab0c68" containerName="registry-server" Oct 01 10:31:41 crc kubenswrapper[4787]: I1001 10:31:41.850998 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4twfr/must-gather-89b86" Oct 01 10:31:41 crc kubenswrapper[4787]: I1001 10:31:41.868162 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-4twfr"/"kube-root-ca.crt" Oct 01 10:31:41 crc kubenswrapper[4787]: I1001 10:31:41.868956 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-4twfr"/"openshift-service-ca.crt" Oct 01 10:31:41 crc kubenswrapper[4787]: I1001 10:31:41.873015 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-4twfr/must-gather-89b86"] Oct 01 10:31:41 crc kubenswrapper[4787]: I1001 10:31:41.989257 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe-must-gather-output\") pod \"must-gather-89b86\" (UID: \"cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe\") " pod="openshift-must-gather-4twfr/must-gather-89b86" Oct 01 10:31:41 crc kubenswrapper[4787]: I1001 10:31:41.989934 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grprh\" (UniqueName: \"kubernetes.io/projected/cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe-kube-api-access-grprh\") pod \"must-gather-89b86\" (UID: \"cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe\") " pod="openshift-must-gather-4twfr/must-gather-89b86" Oct 01 10:31:42 crc kubenswrapper[4787]: I1001 10:31:42.092205 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grprh\" (UniqueName: \"kubernetes.io/projected/cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe-kube-api-access-grprh\") pod \"must-gather-89b86\" (UID: \"cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe\") " pod="openshift-must-gather-4twfr/must-gather-89b86" Oct 01 10:31:42 crc kubenswrapper[4787]: I1001 10:31:42.092314 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe-must-gather-output\") pod \"must-gather-89b86\" (UID: \"cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe\") " pod="openshift-must-gather-4twfr/must-gather-89b86" Oct 01 10:31:42 crc kubenswrapper[4787]: I1001 10:31:42.092794 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe-must-gather-output\") pod \"must-gather-89b86\" (UID: \"cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe\") " pod="openshift-must-gather-4twfr/must-gather-89b86" Oct 01 10:31:42 crc kubenswrapper[4787]: I1001 10:31:42.116314 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grprh\" (UniqueName: \"kubernetes.io/projected/cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe-kube-api-access-grprh\") pod \"must-gather-89b86\" (UID: \"cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe\") " pod="openshift-must-gather-4twfr/must-gather-89b86" Oct 01 10:31:42 crc kubenswrapper[4787]: I1001 10:31:42.176280 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4twfr/must-gather-89b86" Oct 01 10:31:42 crc kubenswrapper[4787]: I1001 10:31:42.658955 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-4twfr/must-gather-89b86"] Oct 01 10:31:43 crc kubenswrapper[4787]: I1001 10:31:43.136308 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4twfr/must-gather-89b86" event={"ID":"cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe","Type":"ContainerStarted","Data":"05f0ab7ac3f71878156a841884db26e5b1b16713886423b406037261ee986d52"} Oct 01 10:31:47 crc kubenswrapper[4787]: I1001 10:31:47.170892 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4twfr/must-gather-89b86" event={"ID":"cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe","Type":"ContainerStarted","Data":"4aa83081c78d868deba718968a1a161f88c8c2c16bca9fcf3ca5ad6a06dc8700"} Oct 01 10:31:47 crc kubenswrapper[4787]: I1001 10:31:47.171402 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4twfr/must-gather-89b86" event={"ID":"cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe","Type":"ContainerStarted","Data":"b4f7cff751b06a5d5c173da30c66c55eb2db181c7d012a6ab5963c9665355dd4"} Oct 01 10:31:47 crc kubenswrapper[4787]: I1001 10:31:47.194368 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-4twfr/must-gather-89b86" podStartSLOduration=2.706507139 podStartE2EDuration="6.194351648s" podCreationTimestamp="2025-10-01 10:31:41 +0000 UTC" firstStartedPulling="2025-10-01 10:31:42.66048704 +0000 UTC m=+3334.775631197" lastFinishedPulling="2025-10-01 10:31:46.148331549 +0000 UTC m=+3338.263475706" observedRunningTime="2025-10-01 10:31:47.185203694 +0000 UTC m=+3339.300347881" watchObservedRunningTime="2025-10-01 10:31:47.194351648 +0000 UTC m=+3339.309495805" Oct 01 10:31:50 crc kubenswrapper[4787]: I1001 10:31:50.102858 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-4twfr/crc-debug-cmf6s"] Oct 01 10:31:50 crc kubenswrapper[4787]: I1001 10:31:50.105212 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4twfr/crc-debug-cmf6s" Oct 01 10:31:50 crc kubenswrapper[4787]: I1001 10:31:50.108895 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-4twfr"/"default-dockercfg-m79wp" Oct 01 10:31:50 crc kubenswrapper[4787]: I1001 10:31:50.155674 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1f19a4ae-ba42-42dd-8bd3-1e93aadb2620-host\") pod \"crc-debug-cmf6s\" (UID: \"1f19a4ae-ba42-42dd-8bd3-1e93aadb2620\") " pod="openshift-must-gather-4twfr/crc-debug-cmf6s" Oct 01 10:31:50 crc kubenswrapper[4787]: I1001 10:31:50.155821 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqlxd\" (UniqueName: \"kubernetes.io/projected/1f19a4ae-ba42-42dd-8bd3-1e93aadb2620-kube-api-access-cqlxd\") pod \"crc-debug-cmf6s\" (UID: \"1f19a4ae-ba42-42dd-8bd3-1e93aadb2620\") " pod="openshift-must-gather-4twfr/crc-debug-cmf6s" Oct 01 10:31:50 crc kubenswrapper[4787]: I1001 10:31:50.259571 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1f19a4ae-ba42-42dd-8bd3-1e93aadb2620-host\") pod \"crc-debug-cmf6s\" (UID: \"1f19a4ae-ba42-42dd-8bd3-1e93aadb2620\") " pod="openshift-must-gather-4twfr/crc-debug-cmf6s" Oct 01 10:31:50 crc kubenswrapper[4787]: I1001 10:31:50.259879 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqlxd\" (UniqueName: \"kubernetes.io/projected/1f19a4ae-ba42-42dd-8bd3-1e93aadb2620-kube-api-access-cqlxd\") pod \"crc-debug-cmf6s\" (UID: \"1f19a4ae-ba42-42dd-8bd3-1e93aadb2620\") " pod="openshift-must-gather-4twfr/crc-debug-cmf6s" Oct 01 10:31:50 crc kubenswrapper[4787]: I1001 10:31:50.260173 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1f19a4ae-ba42-42dd-8bd3-1e93aadb2620-host\") pod \"crc-debug-cmf6s\" (UID: \"1f19a4ae-ba42-42dd-8bd3-1e93aadb2620\") " pod="openshift-must-gather-4twfr/crc-debug-cmf6s" Oct 01 10:31:50 crc kubenswrapper[4787]: I1001 10:31:50.293364 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqlxd\" (UniqueName: \"kubernetes.io/projected/1f19a4ae-ba42-42dd-8bd3-1e93aadb2620-kube-api-access-cqlxd\") pod \"crc-debug-cmf6s\" (UID: \"1f19a4ae-ba42-42dd-8bd3-1e93aadb2620\") " pod="openshift-must-gather-4twfr/crc-debug-cmf6s" Oct 01 10:31:50 crc kubenswrapper[4787]: I1001 10:31:50.432620 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4twfr/crc-debug-cmf6s" Oct 01 10:31:50 crc kubenswrapper[4787]: W1001 10:31:50.475652 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1f19a4ae_ba42_42dd_8bd3_1e93aadb2620.slice/crio-a25d747efced464b288ea0277e2b1db16a426537a75983cac877388526f61b31 WatchSource:0}: Error finding container a25d747efced464b288ea0277e2b1db16a426537a75983cac877388526f61b31: Status 404 returned error can't find the container with id a25d747efced464b288ea0277e2b1db16a426537a75983cac877388526f61b31 Oct 01 10:31:51 crc kubenswrapper[4787]: I1001 10:31:51.209440 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4twfr/crc-debug-cmf6s" event={"ID":"1f19a4ae-ba42-42dd-8bd3-1e93aadb2620","Type":"ContainerStarted","Data":"a25d747efced464b288ea0277e2b1db16a426537a75983cac877388526f61b31"} Oct 01 10:32:04 crc kubenswrapper[4787]: I1001 10:32:04.359225 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4twfr/crc-debug-cmf6s" event={"ID":"1f19a4ae-ba42-42dd-8bd3-1e93aadb2620","Type":"ContainerStarted","Data":"2859f8091b6b21aa4fd604191070c2dd5f6ef835756eb133ca865d2fbda9bce8"} Oct 01 10:32:11 crc kubenswrapper[4787]: I1001 10:32:11.250725 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:32:11 crc kubenswrapper[4787]: I1001 10:32:11.251337 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:32:11 crc kubenswrapper[4787]: I1001 10:32:11.251384 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" Oct 01 10:32:11 crc kubenswrapper[4787]: I1001 10:32:11.252154 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"600ca70fa1dc39a1299d76f2c60704fbc1ec67b77ac9063a4f3e5239fa4265de"} pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 10:32:11 crc kubenswrapper[4787]: I1001 10:32:11.252214 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" containerID="cri-o://600ca70fa1dc39a1299d76f2c60704fbc1ec67b77ac9063a4f3e5239fa4265de" gracePeriod=600 Oct 01 10:32:11 crc kubenswrapper[4787]: I1001 10:32:11.421207 4787 generic.go:334] "Generic (PLEG): container finished" podID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerID="600ca70fa1dc39a1299d76f2c60704fbc1ec67b77ac9063a4f3e5239fa4265de" exitCode=0 Oct 01 10:32:11 crc kubenswrapper[4787]: I1001 10:32:11.421331 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" event={"ID":"fa51b883-5f3e-4141-9d4c-37704aac7718","Type":"ContainerDied","Data":"600ca70fa1dc39a1299d76f2c60704fbc1ec67b77ac9063a4f3e5239fa4265de"} Oct 01 10:32:11 crc kubenswrapper[4787]: I1001 10:32:11.421520 4787 scope.go:117] "RemoveContainer" containerID="3df952eaa93a8322715e6a8d23a95eedf91662575215810014140a3e7d2a03bd" Oct 01 10:32:12 crc kubenswrapper[4787]: I1001 10:32:12.033146 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-4twfr/crc-debug-cmf6s" podStartSLOduration=9.247387833 podStartE2EDuration="22.033124739s" podCreationTimestamp="2025-10-01 10:31:50 +0000 UTC" firstStartedPulling="2025-10-01 10:31:50.478052053 +0000 UTC m=+3342.593196210" lastFinishedPulling="2025-10-01 10:32:03.263788959 +0000 UTC m=+3355.378933116" observedRunningTime="2025-10-01 10:32:04.38909042 +0000 UTC m=+3356.504234577" watchObservedRunningTime="2025-10-01 10:32:12.033124739 +0000 UTC m=+3364.148268906" Oct 01 10:32:12 crc kubenswrapper[4787]: I1001 10:32:12.038312 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bbtx7"] Oct 01 10:32:12 crc kubenswrapper[4787]: I1001 10:32:12.040471 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bbtx7" Oct 01 10:32:12 crc kubenswrapper[4787]: I1001 10:32:12.050888 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bbtx7"] Oct 01 10:32:12 crc kubenswrapper[4787]: I1001 10:32:12.121275 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e07b850-736d-4309-a86f-b8ebf97c0f24-catalog-content\") pod \"redhat-operators-bbtx7\" (UID: \"9e07b850-736d-4309-a86f-b8ebf97c0f24\") " pod="openshift-marketplace/redhat-operators-bbtx7" Oct 01 10:32:12 crc kubenswrapper[4787]: I1001 10:32:12.121370 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79s8c\" (UniqueName: \"kubernetes.io/projected/9e07b850-736d-4309-a86f-b8ebf97c0f24-kube-api-access-79s8c\") pod \"redhat-operators-bbtx7\" (UID: \"9e07b850-736d-4309-a86f-b8ebf97c0f24\") " pod="openshift-marketplace/redhat-operators-bbtx7" Oct 01 10:32:12 crc kubenswrapper[4787]: I1001 10:32:12.121427 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e07b850-736d-4309-a86f-b8ebf97c0f24-utilities\") pod \"redhat-operators-bbtx7\" (UID: \"9e07b850-736d-4309-a86f-b8ebf97c0f24\") " pod="openshift-marketplace/redhat-operators-bbtx7" Oct 01 10:32:12 crc kubenswrapper[4787]: I1001 10:32:12.224403 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79s8c\" (UniqueName: \"kubernetes.io/projected/9e07b850-736d-4309-a86f-b8ebf97c0f24-kube-api-access-79s8c\") pod \"redhat-operators-bbtx7\" (UID: \"9e07b850-736d-4309-a86f-b8ebf97c0f24\") " pod="openshift-marketplace/redhat-operators-bbtx7" Oct 01 10:32:12 crc kubenswrapper[4787]: I1001 10:32:12.224520 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e07b850-736d-4309-a86f-b8ebf97c0f24-utilities\") pod \"redhat-operators-bbtx7\" (UID: \"9e07b850-736d-4309-a86f-b8ebf97c0f24\") " pod="openshift-marketplace/redhat-operators-bbtx7" Oct 01 10:32:12 crc kubenswrapper[4787]: I1001 10:32:12.224699 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e07b850-736d-4309-a86f-b8ebf97c0f24-catalog-content\") pod \"redhat-operators-bbtx7\" (UID: \"9e07b850-736d-4309-a86f-b8ebf97c0f24\") " pod="openshift-marketplace/redhat-operators-bbtx7" Oct 01 10:32:12 crc kubenswrapper[4787]: I1001 10:32:12.225068 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e07b850-736d-4309-a86f-b8ebf97c0f24-utilities\") pod \"redhat-operators-bbtx7\" (UID: \"9e07b850-736d-4309-a86f-b8ebf97c0f24\") " pod="openshift-marketplace/redhat-operators-bbtx7" Oct 01 10:32:12 crc kubenswrapper[4787]: I1001 10:32:12.225305 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e07b850-736d-4309-a86f-b8ebf97c0f24-catalog-content\") pod \"redhat-operators-bbtx7\" (UID: \"9e07b850-736d-4309-a86f-b8ebf97c0f24\") " pod="openshift-marketplace/redhat-operators-bbtx7" Oct 01 10:32:12 crc kubenswrapper[4787]: I1001 10:32:12.257422 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79s8c\" (UniqueName: \"kubernetes.io/projected/9e07b850-736d-4309-a86f-b8ebf97c0f24-kube-api-access-79s8c\") pod \"redhat-operators-bbtx7\" (UID: \"9e07b850-736d-4309-a86f-b8ebf97c0f24\") " pod="openshift-marketplace/redhat-operators-bbtx7" Oct 01 10:32:12 crc kubenswrapper[4787]: I1001 10:32:12.366943 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bbtx7" Oct 01 10:32:12 crc kubenswrapper[4787]: I1001 10:32:12.434952 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" event={"ID":"fa51b883-5f3e-4141-9d4c-37704aac7718","Type":"ContainerStarted","Data":"7cb2890ef30bbede4935afca82be38385def845a6f5b1c0bd5934ac785342b56"} Oct 01 10:32:12 crc kubenswrapper[4787]: I1001 10:32:12.881261 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bbtx7"] Oct 01 10:32:12 crc kubenswrapper[4787]: W1001 10:32:12.894637 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e07b850_736d_4309_a86f_b8ebf97c0f24.slice/crio-0daf51713486e0baa82786cefe400d58c70724bc4070f7a1be71ba1bc3976af6 WatchSource:0}: Error finding container 0daf51713486e0baa82786cefe400d58c70724bc4070f7a1be71ba1bc3976af6: Status 404 returned error can't find the container with id 0daf51713486e0baa82786cefe400d58c70724bc4070f7a1be71ba1bc3976af6 Oct 01 10:32:13 crc kubenswrapper[4787]: I1001 10:32:13.446952 4787 generic.go:334] "Generic (PLEG): container finished" podID="9e07b850-736d-4309-a86f-b8ebf97c0f24" containerID="5ffc0aea4dc94eaee34c221354a95f51f693ce72a34bffbd1e504170eff07b2c" exitCode=0 Oct 01 10:32:13 crc kubenswrapper[4787]: I1001 10:32:13.447041 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bbtx7" event={"ID":"9e07b850-736d-4309-a86f-b8ebf97c0f24","Type":"ContainerDied","Data":"5ffc0aea4dc94eaee34c221354a95f51f693ce72a34bffbd1e504170eff07b2c"} Oct 01 10:32:13 crc kubenswrapper[4787]: I1001 10:32:13.447333 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bbtx7" event={"ID":"9e07b850-736d-4309-a86f-b8ebf97c0f24","Type":"ContainerStarted","Data":"0daf51713486e0baa82786cefe400d58c70724bc4070f7a1be71ba1bc3976af6"} Oct 01 10:32:15 crc kubenswrapper[4787]: I1001 10:32:15.465010 4787 generic.go:334] "Generic (PLEG): container finished" podID="9e07b850-736d-4309-a86f-b8ebf97c0f24" containerID="b96e449b10ec4f22581bd5b7bc127f76572d22ee11e276fcdf8f7dac1a1cc396" exitCode=0 Oct 01 10:32:15 crc kubenswrapper[4787]: I1001 10:32:15.465137 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bbtx7" event={"ID":"9e07b850-736d-4309-a86f-b8ebf97c0f24","Type":"ContainerDied","Data":"b96e449b10ec4f22581bd5b7bc127f76572d22ee11e276fcdf8f7dac1a1cc396"} Oct 01 10:32:19 crc kubenswrapper[4787]: I1001 10:32:19.502905 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bbtx7" event={"ID":"9e07b850-736d-4309-a86f-b8ebf97c0f24","Type":"ContainerStarted","Data":"6c629c5adff5c082a9bccbba0056da821f91c011e3bbc059add71d2ff4ef9cb4"} Oct 01 10:32:22 crc kubenswrapper[4787]: I1001 10:32:22.367373 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bbtx7" Oct 01 10:32:22 crc kubenswrapper[4787]: I1001 10:32:22.367789 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bbtx7" Oct 01 10:32:23 crc kubenswrapper[4787]: I1001 10:32:23.420449 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bbtx7" podUID="9e07b850-736d-4309-a86f-b8ebf97c0f24" containerName="registry-server" probeResult="failure" output=< Oct 01 10:32:23 crc kubenswrapper[4787]: timeout: failed to connect service ":50051" within 1s Oct 01 10:32:23 crc kubenswrapper[4787]: > Oct 01 10:32:32 crc kubenswrapper[4787]: I1001 10:32:32.427686 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bbtx7" Oct 01 10:32:32 crc kubenswrapper[4787]: I1001 10:32:32.463441 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bbtx7" podStartSLOduration=17.40573284 podStartE2EDuration="20.463414826s" podCreationTimestamp="2025-10-01 10:32:12 +0000 UTC" firstStartedPulling="2025-10-01 10:32:13.448765607 +0000 UTC m=+3365.563909764" lastFinishedPulling="2025-10-01 10:32:16.506447583 +0000 UTC m=+3368.621591750" observedRunningTime="2025-10-01 10:32:19.524123947 +0000 UTC m=+3371.639268114" watchObservedRunningTime="2025-10-01 10:32:32.463414826 +0000 UTC m=+3384.578558983" Oct 01 10:32:32 crc kubenswrapper[4787]: I1001 10:32:32.490403 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bbtx7" Oct 01 10:32:32 crc kubenswrapper[4787]: I1001 10:32:32.679353 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bbtx7"] Oct 01 10:32:33 crc kubenswrapper[4787]: I1001 10:32:33.648260 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bbtx7" podUID="9e07b850-736d-4309-a86f-b8ebf97c0f24" containerName="registry-server" containerID="cri-o://6c629c5adff5c082a9bccbba0056da821f91c011e3bbc059add71d2ff4ef9cb4" gracePeriod=2 Oct 01 10:32:34 crc kubenswrapper[4787]: I1001 10:32:34.164463 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bbtx7" Oct 01 10:32:34 crc kubenswrapper[4787]: I1001 10:32:34.250427 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e07b850-736d-4309-a86f-b8ebf97c0f24-utilities\") pod \"9e07b850-736d-4309-a86f-b8ebf97c0f24\" (UID: \"9e07b850-736d-4309-a86f-b8ebf97c0f24\") " Oct 01 10:32:34 crc kubenswrapper[4787]: I1001 10:32:34.250510 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e07b850-736d-4309-a86f-b8ebf97c0f24-catalog-content\") pod \"9e07b850-736d-4309-a86f-b8ebf97c0f24\" (UID: \"9e07b850-736d-4309-a86f-b8ebf97c0f24\") " Oct 01 10:32:34 crc kubenswrapper[4787]: I1001 10:32:34.250589 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79s8c\" (UniqueName: \"kubernetes.io/projected/9e07b850-736d-4309-a86f-b8ebf97c0f24-kube-api-access-79s8c\") pod \"9e07b850-736d-4309-a86f-b8ebf97c0f24\" (UID: \"9e07b850-736d-4309-a86f-b8ebf97c0f24\") " Oct 01 10:32:34 crc kubenswrapper[4787]: I1001 10:32:34.251328 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e07b850-736d-4309-a86f-b8ebf97c0f24-utilities" (OuterVolumeSpecName: "utilities") pod "9e07b850-736d-4309-a86f-b8ebf97c0f24" (UID: "9e07b850-736d-4309-a86f-b8ebf97c0f24"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:32:34 crc kubenswrapper[4787]: I1001 10:32:34.257602 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e07b850-736d-4309-a86f-b8ebf97c0f24-kube-api-access-79s8c" (OuterVolumeSpecName: "kube-api-access-79s8c") pod "9e07b850-736d-4309-a86f-b8ebf97c0f24" (UID: "9e07b850-736d-4309-a86f-b8ebf97c0f24"). InnerVolumeSpecName "kube-api-access-79s8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:32:34 crc kubenswrapper[4787]: I1001 10:32:34.331650 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e07b850-736d-4309-a86f-b8ebf97c0f24-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9e07b850-736d-4309-a86f-b8ebf97c0f24" (UID: "9e07b850-736d-4309-a86f-b8ebf97c0f24"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:32:34 crc kubenswrapper[4787]: I1001 10:32:34.353283 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e07b850-736d-4309-a86f-b8ebf97c0f24-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 10:32:34 crc kubenswrapper[4787]: I1001 10:32:34.353323 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e07b850-736d-4309-a86f-b8ebf97c0f24-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 10:32:34 crc kubenswrapper[4787]: I1001 10:32:34.353340 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79s8c\" (UniqueName: \"kubernetes.io/projected/9e07b850-736d-4309-a86f-b8ebf97c0f24-kube-api-access-79s8c\") on node \"crc\" DevicePath \"\"" Oct 01 10:32:34 crc kubenswrapper[4787]: I1001 10:32:34.669620 4787 generic.go:334] "Generic (PLEG): container finished" podID="9e07b850-736d-4309-a86f-b8ebf97c0f24" containerID="6c629c5adff5c082a9bccbba0056da821f91c011e3bbc059add71d2ff4ef9cb4" exitCode=0 Oct 01 10:32:34 crc kubenswrapper[4787]: I1001 10:32:34.669792 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bbtx7" event={"ID":"9e07b850-736d-4309-a86f-b8ebf97c0f24","Type":"ContainerDied","Data":"6c629c5adff5c082a9bccbba0056da821f91c011e3bbc059add71d2ff4ef9cb4"} Oct 01 10:32:34 crc kubenswrapper[4787]: I1001 10:32:34.670027 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bbtx7" event={"ID":"9e07b850-736d-4309-a86f-b8ebf97c0f24","Type":"ContainerDied","Data":"0daf51713486e0baa82786cefe400d58c70724bc4070f7a1be71ba1bc3976af6"} Oct 01 10:32:34 crc kubenswrapper[4787]: I1001 10:32:34.670054 4787 scope.go:117] "RemoveContainer" containerID="6c629c5adff5c082a9bccbba0056da821f91c011e3bbc059add71d2ff4ef9cb4" Oct 01 10:32:34 crc kubenswrapper[4787]: I1001 10:32:34.669843 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bbtx7" Oct 01 10:32:34 crc kubenswrapper[4787]: I1001 10:32:34.695841 4787 scope.go:117] "RemoveContainer" containerID="b96e449b10ec4f22581bd5b7bc127f76572d22ee11e276fcdf8f7dac1a1cc396" Oct 01 10:32:34 crc kubenswrapper[4787]: I1001 10:32:34.702565 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bbtx7"] Oct 01 10:32:34 crc kubenswrapper[4787]: I1001 10:32:34.709779 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bbtx7"] Oct 01 10:32:34 crc kubenswrapper[4787]: I1001 10:32:34.719763 4787 scope.go:117] "RemoveContainer" containerID="5ffc0aea4dc94eaee34c221354a95f51f693ce72a34bffbd1e504170eff07b2c" Oct 01 10:32:34 crc kubenswrapper[4787]: I1001 10:32:34.811507 4787 scope.go:117] "RemoveContainer" containerID="6c629c5adff5c082a9bccbba0056da821f91c011e3bbc059add71d2ff4ef9cb4" Oct 01 10:32:34 crc kubenswrapper[4787]: E1001 10:32:34.812036 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c629c5adff5c082a9bccbba0056da821f91c011e3bbc059add71d2ff4ef9cb4\": container with ID starting with 6c629c5adff5c082a9bccbba0056da821f91c011e3bbc059add71d2ff4ef9cb4 not found: ID does not exist" containerID="6c629c5adff5c082a9bccbba0056da821f91c011e3bbc059add71d2ff4ef9cb4" Oct 01 10:32:34 crc kubenswrapper[4787]: I1001 10:32:34.812095 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c629c5adff5c082a9bccbba0056da821f91c011e3bbc059add71d2ff4ef9cb4"} err="failed to get container status \"6c629c5adff5c082a9bccbba0056da821f91c011e3bbc059add71d2ff4ef9cb4\": rpc error: code = NotFound desc = could not find container \"6c629c5adff5c082a9bccbba0056da821f91c011e3bbc059add71d2ff4ef9cb4\": container with ID starting with 6c629c5adff5c082a9bccbba0056da821f91c011e3bbc059add71d2ff4ef9cb4 not found: ID does not exist" Oct 01 10:32:34 crc kubenswrapper[4787]: I1001 10:32:34.812135 4787 scope.go:117] "RemoveContainer" containerID="b96e449b10ec4f22581bd5b7bc127f76572d22ee11e276fcdf8f7dac1a1cc396" Oct 01 10:32:34 crc kubenswrapper[4787]: E1001 10:32:34.812385 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b96e449b10ec4f22581bd5b7bc127f76572d22ee11e276fcdf8f7dac1a1cc396\": container with ID starting with b96e449b10ec4f22581bd5b7bc127f76572d22ee11e276fcdf8f7dac1a1cc396 not found: ID does not exist" containerID="b96e449b10ec4f22581bd5b7bc127f76572d22ee11e276fcdf8f7dac1a1cc396" Oct 01 10:32:34 crc kubenswrapper[4787]: I1001 10:32:34.812412 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b96e449b10ec4f22581bd5b7bc127f76572d22ee11e276fcdf8f7dac1a1cc396"} err="failed to get container status \"b96e449b10ec4f22581bd5b7bc127f76572d22ee11e276fcdf8f7dac1a1cc396\": rpc error: code = NotFound desc = could not find container \"b96e449b10ec4f22581bd5b7bc127f76572d22ee11e276fcdf8f7dac1a1cc396\": container with ID starting with b96e449b10ec4f22581bd5b7bc127f76572d22ee11e276fcdf8f7dac1a1cc396 not found: ID does not exist" Oct 01 10:32:34 crc kubenswrapper[4787]: I1001 10:32:34.812429 4787 scope.go:117] "RemoveContainer" containerID="5ffc0aea4dc94eaee34c221354a95f51f693ce72a34bffbd1e504170eff07b2c" Oct 01 10:32:34 crc kubenswrapper[4787]: E1001 10:32:34.812653 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ffc0aea4dc94eaee34c221354a95f51f693ce72a34bffbd1e504170eff07b2c\": container with ID starting with 5ffc0aea4dc94eaee34c221354a95f51f693ce72a34bffbd1e504170eff07b2c not found: ID does not exist" containerID="5ffc0aea4dc94eaee34c221354a95f51f693ce72a34bffbd1e504170eff07b2c" Oct 01 10:32:34 crc kubenswrapper[4787]: I1001 10:32:34.812685 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ffc0aea4dc94eaee34c221354a95f51f693ce72a34bffbd1e504170eff07b2c"} err="failed to get container status \"5ffc0aea4dc94eaee34c221354a95f51f693ce72a34bffbd1e504170eff07b2c\": rpc error: code = NotFound desc = could not find container \"5ffc0aea4dc94eaee34c221354a95f51f693ce72a34bffbd1e504170eff07b2c\": container with ID starting with 5ffc0aea4dc94eaee34c221354a95f51f693ce72a34bffbd1e504170eff07b2c not found: ID does not exist" Oct 01 10:32:36 crc kubenswrapper[4787]: I1001 10:32:36.536241 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e07b850-736d-4309-a86f-b8ebf97c0f24" path="/var/lib/kubelet/pods/9e07b850-736d-4309-a86f-b8ebf97c0f24/volumes" Oct 01 10:32:40 crc kubenswrapper[4787]: I1001 10:32:40.191266 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7tmn2"] Oct 01 10:32:40 crc kubenswrapper[4787]: E1001 10:32:40.192378 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e07b850-736d-4309-a86f-b8ebf97c0f24" containerName="extract-content" Oct 01 10:32:40 crc kubenswrapper[4787]: I1001 10:32:40.192399 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e07b850-736d-4309-a86f-b8ebf97c0f24" containerName="extract-content" Oct 01 10:32:40 crc kubenswrapper[4787]: E1001 10:32:40.192429 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e07b850-736d-4309-a86f-b8ebf97c0f24" containerName="extract-utilities" Oct 01 10:32:40 crc kubenswrapper[4787]: I1001 10:32:40.192439 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e07b850-736d-4309-a86f-b8ebf97c0f24" containerName="extract-utilities" Oct 01 10:32:40 crc kubenswrapper[4787]: E1001 10:32:40.192461 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e07b850-736d-4309-a86f-b8ebf97c0f24" containerName="registry-server" Oct 01 10:32:40 crc kubenswrapper[4787]: I1001 10:32:40.192469 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e07b850-736d-4309-a86f-b8ebf97c0f24" containerName="registry-server" Oct 01 10:32:40 crc kubenswrapper[4787]: I1001 10:32:40.192788 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e07b850-736d-4309-a86f-b8ebf97c0f24" containerName="registry-server" Oct 01 10:32:40 crc kubenswrapper[4787]: I1001 10:32:40.194622 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7tmn2" Oct 01 10:32:40 crc kubenswrapper[4787]: I1001 10:32:40.217124 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7tmn2"] Oct 01 10:32:40 crc kubenswrapper[4787]: I1001 10:32:40.277671 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e-utilities\") pod \"redhat-marketplace-7tmn2\" (UID: \"2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e\") " pod="openshift-marketplace/redhat-marketplace-7tmn2" Oct 01 10:32:40 crc kubenswrapper[4787]: I1001 10:32:40.277733 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e-catalog-content\") pod \"redhat-marketplace-7tmn2\" (UID: \"2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e\") " pod="openshift-marketplace/redhat-marketplace-7tmn2" Oct 01 10:32:40 crc kubenswrapper[4787]: I1001 10:32:40.277872 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgmm5\" (UniqueName: \"kubernetes.io/projected/2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e-kube-api-access-qgmm5\") pod \"redhat-marketplace-7tmn2\" (UID: \"2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e\") " pod="openshift-marketplace/redhat-marketplace-7tmn2" Oct 01 10:32:40 crc kubenswrapper[4787]: I1001 10:32:40.379396 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e-utilities\") pod \"redhat-marketplace-7tmn2\" (UID: \"2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e\") " pod="openshift-marketplace/redhat-marketplace-7tmn2" Oct 01 10:32:40 crc kubenswrapper[4787]: I1001 10:32:40.379448 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e-catalog-content\") pod \"redhat-marketplace-7tmn2\" (UID: \"2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e\") " pod="openshift-marketplace/redhat-marketplace-7tmn2" Oct 01 10:32:40 crc kubenswrapper[4787]: I1001 10:32:40.379586 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgmm5\" (UniqueName: \"kubernetes.io/projected/2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e-kube-api-access-qgmm5\") pod \"redhat-marketplace-7tmn2\" (UID: \"2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e\") " pod="openshift-marketplace/redhat-marketplace-7tmn2" Oct 01 10:32:40 crc kubenswrapper[4787]: I1001 10:32:40.380128 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e-utilities\") pod \"redhat-marketplace-7tmn2\" (UID: \"2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e\") " pod="openshift-marketplace/redhat-marketplace-7tmn2" Oct 01 10:32:40 crc kubenswrapper[4787]: I1001 10:32:40.380389 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e-catalog-content\") pod \"redhat-marketplace-7tmn2\" (UID: \"2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e\") " pod="openshift-marketplace/redhat-marketplace-7tmn2" Oct 01 10:32:40 crc kubenswrapper[4787]: I1001 10:32:40.403686 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgmm5\" (UniqueName: \"kubernetes.io/projected/2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e-kube-api-access-qgmm5\") pod \"redhat-marketplace-7tmn2\" (UID: \"2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e\") " pod="openshift-marketplace/redhat-marketplace-7tmn2" Oct 01 10:32:40 crc kubenswrapper[4787]: I1001 10:32:40.532750 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7tmn2" Oct 01 10:32:41 crc kubenswrapper[4787]: I1001 10:32:41.043744 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7tmn2"] Oct 01 10:32:41 crc kubenswrapper[4787]: I1001 10:32:41.758970 4787 generic.go:334] "Generic (PLEG): container finished" podID="2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e" containerID="0cdf12aa686915ddc4c25001bdab2acc96c983ab0b9c1d4338497009fc6c19fe" exitCode=0 Oct 01 10:32:41 crc kubenswrapper[4787]: I1001 10:32:41.759547 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7tmn2" event={"ID":"2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e","Type":"ContainerDied","Data":"0cdf12aa686915ddc4c25001bdab2acc96c983ab0b9c1d4338497009fc6c19fe"} Oct 01 10:32:41 crc kubenswrapper[4787]: I1001 10:32:41.761835 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7tmn2" event={"ID":"2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e","Type":"ContainerStarted","Data":"7dca6740a3a7ac9ed710125d30149e47272e2b8be7e5bec56f80841348f59ffe"} Oct 01 10:32:43 crc kubenswrapper[4787]: I1001 10:32:43.784545 4787 generic.go:334] "Generic (PLEG): container finished" podID="2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e" containerID="4f0a9825253eadf29a886e4fe841db8aba52c5fc20c0bd6140c280d0b73d80e7" exitCode=0 Oct 01 10:32:43 crc kubenswrapper[4787]: I1001 10:32:43.784679 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7tmn2" event={"ID":"2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e","Type":"ContainerDied","Data":"4f0a9825253eadf29a886e4fe841db8aba52c5fc20c0bd6140c280d0b73d80e7"} Oct 01 10:32:44 crc kubenswrapper[4787]: I1001 10:32:44.833134 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7tmn2" event={"ID":"2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e","Type":"ContainerStarted","Data":"ea1636883a8c85ea46fb39595fc47ff09fa281cbf3bade7d58de98c40af786a4"} Oct 01 10:32:44 crc kubenswrapper[4787]: I1001 10:32:44.854789 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7tmn2" podStartSLOduration=2.375730783 podStartE2EDuration="4.854771165s" podCreationTimestamp="2025-10-01 10:32:40 +0000 UTC" firstStartedPulling="2025-10-01 10:32:41.760991666 +0000 UTC m=+3393.876135823" lastFinishedPulling="2025-10-01 10:32:44.240032058 +0000 UTC m=+3396.355176205" observedRunningTime="2025-10-01 10:32:44.854510069 +0000 UTC m=+3396.969654236" watchObservedRunningTime="2025-10-01 10:32:44.854771165 +0000 UTC m=+3396.969915312" Oct 01 10:32:50 crc kubenswrapper[4787]: I1001 10:32:50.535864 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7tmn2" Oct 01 10:32:50 crc kubenswrapper[4787]: I1001 10:32:50.536438 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7tmn2" Oct 01 10:32:50 crc kubenswrapper[4787]: I1001 10:32:50.592043 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7tmn2" Oct 01 10:32:50 crc kubenswrapper[4787]: I1001 10:32:50.942036 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7tmn2" Oct 01 10:32:50 crc kubenswrapper[4787]: I1001 10:32:50.990525 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7tmn2"] Oct 01 10:32:52 crc kubenswrapper[4787]: I1001 10:32:52.909304 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7tmn2" podUID="2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e" containerName="registry-server" containerID="cri-o://ea1636883a8c85ea46fb39595fc47ff09fa281cbf3bade7d58de98c40af786a4" gracePeriod=2 Oct 01 10:32:53 crc kubenswrapper[4787]: I1001 10:32:53.400890 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7tmn2" Oct 01 10:32:53 crc kubenswrapper[4787]: I1001 10:32:53.559572 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e-catalog-content\") pod \"2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e\" (UID: \"2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e\") " Oct 01 10:32:53 crc kubenswrapper[4787]: I1001 10:32:53.560044 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgmm5\" (UniqueName: \"kubernetes.io/projected/2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e-kube-api-access-qgmm5\") pod \"2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e\" (UID: \"2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e\") " Oct 01 10:32:53 crc kubenswrapper[4787]: I1001 10:32:53.560683 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e-utilities\") pod \"2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e\" (UID: \"2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e\") " Oct 01 10:32:53 crc kubenswrapper[4787]: I1001 10:32:53.561413 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e-utilities" (OuterVolumeSpecName: "utilities") pod "2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e" (UID: "2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:32:53 crc kubenswrapper[4787]: I1001 10:32:53.566483 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e-kube-api-access-qgmm5" (OuterVolumeSpecName: "kube-api-access-qgmm5") pod "2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e" (UID: "2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e"). InnerVolumeSpecName "kube-api-access-qgmm5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:32:53 crc kubenswrapper[4787]: I1001 10:32:53.582154 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e" (UID: "2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:32:53 crc kubenswrapper[4787]: I1001 10:32:53.665507 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgmm5\" (UniqueName: \"kubernetes.io/projected/2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e-kube-api-access-qgmm5\") on node \"crc\" DevicePath \"\"" Oct 01 10:32:53 crc kubenswrapper[4787]: I1001 10:32:53.665543 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 10:32:53 crc kubenswrapper[4787]: I1001 10:32:53.665555 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 10:32:53 crc kubenswrapper[4787]: I1001 10:32:53.921792 4787 generic.go:334] "Generic (PLEG): container finished" podID="2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e" containerID="ea1636883a8c85ea46fb39595fc47ff09fa281cbf3bade7d58de98c40af786a4" exitCode=0 Oct 01 10:32:53 crc kubenswrapper[4787]: I1001 10:32:53.921843 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7tmn2" event={"ID":"2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e","Type":"ContainerDied","Data":"ea1636883a8c85ea46fb39595fc47ff09fa281cbf3bade7d58de98c40af786a4"} Oct 01 10:32:53 crc kubenswrapper[4787]: I1001 10:32:53.921872 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7tmn2" event={"ID":"2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e","Type":"ContainerDied","Data":"7dca6740a3a7ac9ed710125d30149e47272e2b8be7e5bec56f80841348f59ffe"} Oct 01 10:32:53 crc kubenswrapper[4787]: I1001 10:32:53.921892 4787 scope.go:117] "RemoveContainer" containerID="ea1636883a8c85ea46fb39595fc47ff09fa281cbf3bade7d58de98c40af786a4" Oct 01 10:32:53 crc kubenswrapper[4787]: I1001 10:32:53.921913 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7tmn2" Oct 01 10:32:53 crc kubenswrapper[4787]: I1001 10:32:53.949047 4787 scope.go:117] "RemoveContainer" containerID="4f0a9825253eadf29a886e4fe841db8aba52c5fc20c0bd6140c280d0b73d80e7" Oct 01 10:32:53 crc kubenswrapper[4787]: I1001 10:32:53.969955 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7tmn2"] Oct 01 10:32:53 crc kubenswrapper[4787]: I1001 10:32:53.978500 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7tmn2"] Oct 01 10:32:53 crc kubenswrapper[4787]: I1001 10:32:53.988646 4787 scope.go:117] "RemoveContainer" containerID="0cdf12aa686915ddc4c25001bdab2acc96c983ab0b9c1d4338497009fc6c19fe" Oct 01 10:32:54 crc kubenswrapper[4787]: I1001 10:32:54.083844 4787 scope.go:117] "RemoveContainer" containerID="ea1636883a8c85ea46fb39595fc47ff09fa281cbf3bade7d58de98c40af786a4" Oct 01 10:32:54 crc kubenswrapper[4787]: E1001 10:32:54.088186 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea1636883a8c85ea46fb39595fc47ff09fa281cbf3bade7d58de98c40af786a4\": container with ID starting with ea1636883a8c85ea46fb39595fc47ff09fa281cbf3bade7d58de98c40af786a4 not found: ID does not exist" containerID="ea1636883a8c85ea46fb39595fc47ff09fa281cbf3bade7d58de98c40af786a4" Oct 01 10:32:54 crc kubenswrapper[4787]: I1001 10:32:54.088279 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea1636883a8c85ea46fb39595fc47ff09fa281cbf3bade7d58de98c40af786a4"} err="failed to get container status \"ea1636883a8c85ea46fb39595fc47ff09fa281cbf3bade7d58de98c40af786a4\": rpc error: code = NotFound desc = could not find container \"ea1636883a8c85ea46fb39595fc47ff09fa281cbf3bade7d58de98c40af786a4\": container with ID starting with ea1636883a8c85ea46fb39595fc47ff09fa281cbf3bade7d58de98c40af786a4 not found: ID does not exist" Oct 01 10:32:54 crc kubenswrapper[4787]: I1001 10:32:54.088332 4787 scope.go:117] "RemoveContainer" containerID="4f0a9825253eadf29a886e4fe841db8aba52c5fc20c0bd6140c280d0b73d80e7" Oct 01 10:32:54 crc kubenswrapper[4787]: E1001 10:32:54.097437 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f0a9825253eadf29a886e4fe841db8aba52c5fc20c0bd6140c280d0b73d80e7\": container with ID starting with 4f0a9825253eadf29a886e4fe841db8aba52c5fc20c0bd6140c280d0b73d80e7 not found: ID does not exist" containerID="4f0a9825253eadf29a886e4fe841db8aba52c5fc20c0bd6140c280d0b73d80e7" Oct 01 10:32:54 crc kubenswrapper[4787]: I1001 10:32:54.097517 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f0a9825253eadf29a886e4fe841db8aba52c5fc20c0bd6140c280d0b73d80e7"} err="failed to get container status \"4f0a9825253eadf29a886e4fe841db8aba52c5fc20c0bd6140c280d0b73d80e7\": rpc error: code = NotFound desc = could not find container \"4f0a9825253eadf29a886e4fe841db8aba52c5fc20c0bd6140c280d0b73d80e7\": container with ID starting with 4f0a9825253eadf29a886e4fe841db8aba52c5fc20c0bd6140c280d0b73d80e7 not found: ID does not exist" Oct 01 10:32:54 crc kubenswrapper[4787]: I1001 10:32:54.097559 4787 scope.go:117] "RemoveContainer" containerID="0cdf12aa686915ddc4c25001bdab2acc96c983ab0b9c1d4338497009fc6c19fe" Oct 01 10:32:54 crc kubenswrapper[4787]: E1001 10:32:54.098151 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0cdf12aa686915ddc4c25001bdab2acc96c983ab0b9c1d4338497009fc6c19fe\": container with ID starting with 0cdf12aa686915ddc4c25001bdab2acc96c983ab0b9c1d4338497009fc6c19fe not found: ID does not exist" containerID="0cdf12aa686915ddc4c25001bdab2acc96c983ab0b9c1d4338497009fc6c19fe" Oct 01 10:32:54 crc kubenswrapper[4787]: I1001 10:32:54.098220 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0cdf12aa686915ddc4c25001bdab2acc96c983ab0b9c1d4338497009fc6c19fe"} err="failed to get container status \"0cdf12aa686915ddc4c25001bdab2acc96c983ab0b9c1d4338497009fc6c19fe\": rpc error: code = NotFound desc = could not find container \"0cdf12aa686915ddc4c25001bdab2acc96c983ab0b9c1d4338497009fc6c19fe\": container with ID starting with 0cdf12aa686915ddc4c25001bdab2acc96c983ab0b9c1d4338497009fc6c19fe not found: ID does not exist" Oct 01 10:32:54 crc kubenswrapper[4787]: I1001 10:32:54.535895 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e" path="/var/lib/kubelet/pods/2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e/volumes" Oct 01 10:32:54 crc kubenswrapper[4787]: I1001 10:32:54.785231 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-887c78f44-f74pm_873541c7-e7e7-4219-8ce0-c8d5edfbe308/barbican-api/0.log" Oct 01 10:32:54 crc kubenswrapper[4787]: I1001 10:32:54.855910 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-887c78f44-f74pm_873541c7-e7e7-4219-8ce0-c8d5edfbe308/barbican-api-log/0.log" Oct 01 10:32:55 crc kubenswrapper[4787]: I1001 10:32:55.082281 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-699b4d5c4-jdg79_484a5c7d-7633-4e34-a235-af316bbccada/barbican-keystone-listener-log/0.log" Oct 01 10:32:55 crc kubenswrapper[4787]: I1001 10:32:55.088507 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-699b4d5c4-jdg79_484a5c7d-7633-4e34-a235-af316bbccada/barbican-keystone-listener/0.log" Oct 01 10:32:55 crc kubenswrapper[4787]: I1001 10:32:55.272547 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-64d9c8b8f9-zv7w7_ea69f2cb-9d4f-4b3c-a0ba-8a3fc8fce274/barbican-worker-log/0.log" Oct 01 10:32:55 crc kubenswrapper[4787]: I1001 10:32:55.328349 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-64d9c8b8f9-zv7w7_ea69f2cb-9d4f-4b3c-a0ba-8a3fc8fce274/barbican-worker/0.log" Oct 01 10:32:55 crc kubenswrapper[4787]: I1001 10:32:55.500924 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-s4tf6_d1018ece-ee61-4913-bd2f-64f996ef135c/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 10:32:55 crc kubenswrapper[4787]: I1001 10:32:55.737817 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0cdb08fb-0a31-40dd-98c6-d13926671f45/ceilometer-central-agent/0.log" Oct 01 10:32:55 crc kubenswrapper[4787]: I1001 10:32:55.741592 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0cdb08fb-0a31-40dd-98c6-d13926671f45/ceilometer-notification-agent/0.log" Oct 01 10:32:55 crc kubenswrapper[4787]: I1001 10:32:55.811819 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0cdb08fb-0a31-40dd-98c6-d13926671f45/proxy-httpd/0.log" Oct 01 10:32:55 crc kubenswrapper[4787]: I1001 10:32:55.955028 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0cdb08fb-0a31-40dd-98c6-d13926671f45/sg-core/0.log" Oct 01 10:32:56 crc kubenswrapper[4787]: I1001 10:32:56.092334 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_a05bd18f-4272-4ff1-ac85-65f0a20cbdcf/cinder-api/0.log" Oct 01 10:32:56 crc kubenswrapper[4787]: I1001 10:32:56.225970 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_a05bd18f-4272-4ff1-ac85-65f0a20cbdcf/cinder-api-log/0.log" Oct 01 10:32:56 crc kubenswrapper[4787]: I1001 10:32:56.323864 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_8a844181-c2d3-4755-a0e2-d3706a9eb3e0/cinder-scheduler/0.log" Oct 01 10:32:56 crc kubenswrapper[4787]: I1001 10:32:56.435375 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_8a844181-c2d3-4755-a0e2-d3706a9eb3e0/probe/0.log" Oct 01 10:32:56 crc kubenswrapper[4787]: I1001 10:32:56.540188 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-llpfd_8dd2f969-9904-4848-b0e5-f852a8b958ea/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 10:32:56 crc kubenswrapper[4787]: I1001 10:32:56.756017 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-ddl72_def7d40c-8191-4256-b8ad-4fe9272d5fae/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 10:32:56 crc kubenswrapper[4787]: I1001 10:32:56.917328 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-zwc7x_c2b165b4-e2e9-4b0c-91f7-db75bcbc113d/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 10:32:57 crc kubenswrapper[4787]: I1001 10:32:57.038214 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6966dc867f-cvjjh_fc33ac30-efd1-438d-9ab5-d700dfd27efe/init/0.log" Oct 01 10:32:57 crc kubenswrapper[4787]: I1001 10:32:57.228486 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6966dc867f-cvjjh_fc33ac30-efd1-438d-9ab5-d700dfd27efe/init/0.log" Oct 01 10:32:57 crc kubenswrapper[4787]: I1001 10:32:57.270030 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6966dc867f-cvjjh_fc33ac30-efd1-438d-9ab5-d700dfd27efe/dnsmasq-dns/0.log" Oct 01 10:32:57 crc kubenswrapper[4787]: I1001 10:32:57.460891 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-4jp68_f70ffa29-bf25-44a2-bd45-822e5f59a5d5/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 10:32:57 crc kubenswrapper[4787]: I1001 10:32:57.508023 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_137125b0-e356-4dbc-be53-2dbd3a06825f/glance-httpd/0.log" Oct 01 10:32:57 crc kubenswrapper[4787]: I1001 10:32:57.697568 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_137125b0-e356-4dbc-be53-2dbd3a06825f/glance-log/0.log" Oct 01 10:32:57 crc kubenswrapper[4787]: I1001 10:32:57.723053 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_ddca8826-066e-45d5-9536-59ef0bfc8c61/glance-httpd/0.log" Oct 01 10:32:57 crc kubenswrapper[4787]: I1001 10:32:57.923430 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_ddca8826-066e-45d5-9536-59ef0bfc8c61/glance-log/0.log" Oct 01 10:32:58 crc kubenswrapper[4787]: I1001 10:32:58.105064 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7d65d5957b-44rtv_db1d14ba-8d00-4508-9c89-fdadb562ad0d/horizon/0.log" Oct 01 10:32:58 crc kubenswrapper[4787]: I1001 10:32:58.273121 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-79skv_d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 10:32:58 crc kubenswrapper[4787]: I1001 10:32:58.347399 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7d65d5957b-44rtv_db1d14ba-8d00-4508-9c89-fdadb562ad0d/horizon-log/0.log" Oct 01 10:32:58 crc kubenswrapper[4787]: I1001 10:32:58.512731 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-qvjjs_654cfd15-20ca-4754-9de7-ebe67905c8d2/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 10:32:58 crc kubenswrapper[4787]: I1001 10:32:58.780116 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29321881-dnxcs_7b0f7d9e-ab06-41c1-bac1-54004f737044/keystone-cron/0.log" Oct 01 10:32:58 crc kubenswrapper[4787]: I1001 10:32:58.789212 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-7f57487985-p72bt_db5966fa-e6ce-4fd2-9a2c-b17671db2625/keystone-api/0.log" Oct 01 10:32:58 crc kubenswrapper[4787]: I1001 10:32:58.943522 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_c1c9f3e0-fa6b-4ef4-ae4a-a0aea3bfc262/kube-state-metrics/0.log" Oct 01 10:32:59 crc kubenswrapper[4787]: I1001 10:32:59.020681 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr_c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 10:32:59 crc kubenswrapper[4787]: I1001 10:32:59.423739 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-58d9cd74f7-86sts_1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0/neutron-httpd/0.log" Oct 01 10:32:59 crc kubenswrapper[4787]: I1001 10:32:59.435429 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-58d9cd74f7-86sts_1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0/neutron-api/0.log" Oct 01 10:32:59 crc kubenswrapper[4787]: I1001 10:32:59.701097 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6_e7fb97de-2759-41c1-b090-99ed2c95e92c/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 10:33:00 crc kubenswrapper[4787]: I1001 10:33:00.278913 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_59257c48-7bb8-4aad-9f6c-71d4459f2fb7/nova-api-log/0.log" Oct 01 10:33:00 crc kubenswrapper[4787]: I1001 10:33:00.517546 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_39f3448f-eee4-419b-a33b-8776fc5df599/nova-cell0-conductor-conductor/0.log" Oct 01 10:33:00 crc kubenswrapper[4787]: I1001 10:33:00.546307 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_59257c48-7bb8-4aad-9f6c-71d4459f2fb7/nova-api-api/0.log" Oct 01 10:33:01 crc kubenswrapper[4787]: I1001 10:33:01.140221 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_1f4a94ee-d626-48f8-93c2-547212f23da4/nova-cell1-conductor-conductor/0.log" Oct 01 10:33:01 crc kubenswrapper[4787]: I1001 10:33:01.182851 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_f8b719ce-ceea-4c80-8c8d-846f2bf7a402/nova-cell1-novncproxy-novncproxy/0.log" Oct 01 10:33:01 crc kubenswrapper[4787]: I1001 10:33:01.499198 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-ltfkq_4f22f28b-714f-4ab7-a90b-a400a500e810/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 10:33:01 crc kubenswrapper[4787]: I1001 10:33:01.774246 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f3dac849-090f-4eaa-bd8a-11a9e85f0317/nova-metadata-log/0.log" Oct 01 10:33:02 crc kubenswrapper[4787]: I1001 10:33:02.287053 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_35a5073c-d1ab-4fa9-b1c5-3a6df1d42bcc/nova-scheduler-scheduler/0.log" Oct 01 10:33:02 crc kubenswrapper[4787]: I1001 10:33:02.450252 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7924a2ac-f411-42d6-8b23-f22faf58351a/mysql-bootstrap/0.log" Oct 01 10:33:02 crc kubenswrapper[4787]: I1001 10:33:02.713122 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7924a2ac-f411-42d6-8b23-f22faf58351a/galera/0.log" Oct 01 10:33:02 crc kubenswrapper[4787]: I1001 10:33:02.748172 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7924a2ac-f411-42d6-8b23-f22faf58351a/mysql-bootstrap/0.log" Oct 01 10:33:03 crc kubenswrapper[4787]: I1001 10:33:03.107140 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c/mysql-bootstrap/0.log" Oct 01 10:33:03 crc kubenswrapper[4787]: I1001 10:33:03.176852 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f3dac849-090f-4eaa-bd8a-11a9e85f0317/nova-metadata-metadata/0.log" Oct 01 10:33:03 crc kubenswrapper[4787]: I1001 10:33:03.313581 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c/mysql-bootstrap/0.log" Oct 01 10:33:03 crc kubenswrapper[4787]: I1001 10:33:03.386422 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c/galera/0.log" Oct 01 10:33:03 crc kubenswrapper[4787]: I1001 10:33:03.496608 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_ccb620ed-ad28-4ec7-8319-6d5bb3ce10ec/openstackclient/0.log" Oct 01 10:33:03 crc kubenswrapper[4787]: I1001 10:33:03.966355 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-l8fqd_36eb99a8-609a-4af2-9fa7-d0051806659e/ovn-controller/0.log" Oct 01 10:33:04 crc kubenswrapper[4787]: I1001 10:33:04.191478 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-d9cxp_11dbf4b0-9694-40e2-a122-9da928c50ad8/openstack-network-exporter/0.log" Oct 01 10:33:04 crc kubenswrapper[4787]: I1001 10:33:04.282758 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6htcd_948a10c7-fb61-4110-a08e-b2fd42cb842a/ovsdb-server-init/0.log" Oct 01 10:33:04 crc kubenswrapper[4787]: I1001 10:33:04.587031 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6htcd_948a10c7-fb61-4110-a08e-b2fd42cb842a/ovs-vswitchd/0.log" Oct 01 10:33:04 crc kubenswrapper[4787]: I1001 10:33:04.613425 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6htcd_948a10c7-fb61-4110-a08e-b2fd42cb842a/ovsdb-server-init/0.log" Oct 01 10:33:04 crc kubenswrapper[4787]: I1001 10:33:04.644662 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6htcd_948a10c7-fb61-4110-a08e-b2fd42cb842a/ovsdb-server/0.log" Oct 01 10:33:04 crc kubenswrapper[4787]: I1001 10:33:04.887731 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-htlbc_dc18c70f-f7c7-4a60-87e6-699320d382fc/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 10:33:05 crc kubenswrapper[4787]: I1001 10:33:05.077994 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_e5f4597a-1f01-4082-aada-cc7b976fad9a/openstack-network-exporter/0.log" Oct 01 10:33:05 crc kubenswrapper[4787]: I1001 10:33:05.106537 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_e5f4597a-1f01-4082-aada-cc7b976fad9a/ovn-northd/0.log" Oct 01 10:33:05 crc kubenswrapper[4787]: I1001 10:33:05.367642 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_ee50aeec-8875-45c2-80e2-7121480ae89f/openstack-network-exporter/0.log" Oct 01 10:33:05 crc kubenswrapper[4787]: I1001 10:33:05.376339 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_ee50aeec-8875-45c2-80e2-7121480ae89f/ovsdbserver-nb/0.log" Oct 01 10:33:05 crc kubenswrapper[4787]: I1001 10:33:05.597533 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_696f6d7f-812c-4d36-a799-397b83edf291/openstack-network-exporter/0.log" Oct 01 10:33:05 crc kubenswrapper[4787]: I1001 10:33:05.643957 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_696f6d7f-812c-4d36-a799-397b83edf291/ovsdbserver-sb/0.log" Oct 01 10:33:05 crc kubenswrapper[4787]: I1001 10:33:05.908409 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-84d5f4684d-p2xjn_a617751d-ce49-4357-bed0-32a3f63d4d84/placement-api/0.log" Oct 01 10:33:05 crc kubenswrapper[4787]: I1001 10:33:05.987972 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-84d5f4684d-p2xjn_a617751d-ce49-4357-bed0-32a3f63d4d84/placement-log/0.log" Oct 01 10:33:06 crc kubenswrapper[4787]: I1001 10:33:06.176194 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_d0353483-c978-4d9a-9c34-73ee379325ba/setup-container/0.log" Oct 01 10:33:06 crc kubenswrapper[4787]: I1001 10:33:06.334115 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_d0353483-c978-4d9a-9c34-73ee379325ba/setup-container/0.log" Oct 01 10:33:06 crc kubenswrapper[4787]: I1001 10:33:06.374669 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_d0353483-c978-4d9a-9c34-73ee379325ba/rabbitmq/0.log" Oct 01 10:33:06 crc kubenswrapper[4787]: I1001 10:33:06.530364 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_eb5022f8-276b-4d6d-917b-be96a1d993bf/setup-container/0.log" Oct 01 10:33:06 crc kubenswrapper[4787]: I1001 10:33:06.817327 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_eb5022f8-276b-4d6d-917b-be96a1d993bf/rabbitmq/0.log" Oct 01 10:33:06 crc kubenswrapper[4787]: I1001 10:33:06.878043 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_eb5022f8-276b-4d6d-917b-be96a1d993bf/setup-container/0.log" Oct 01 10:33:07 crc kubenswrapper[4787]: I1001 10:33:07.135716 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-llnrf_728521d2-1e71-4da3-a8bc-bd68e02eae35/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 10:33:07 crc kubenswrapper[4787]: I1001 10:33:07.174993 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-j5ltq_8d46852a-ab64-4e0c-b29a-e9bde8305a94/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 10:33:07 crc kubenswrapper[4787]: I1001 10:33:07.640798 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-wk6cx_d8ba2dbe-8a68-4da2-ac35-664a50fb5061/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 10:33:08 crc kubenswrapper[4787]: I1001 10:33:08.028591 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-ws2cx_389d6832-2d3e-49a0-afbd-f88359db6324/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 10:33:08 crc kubenswrapper[4787]: I1001 10:33:08.093425 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-rhhxm_49e3ae1d-7abd-4e22-a333-54e10db349a6/ssh-known-hosts-edpm-deployment/0.log" Oct 01 10:33:08 crc kubenswrapper[4787]: I1001 10:33:08.389431 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5c7b59dd4f-spxm4_e743cd5e-e8c6-4fe2-9480-3a30316b8e23/proxy-server/0.log" Oct 01 10:33:08 crc kubenswrapper[4787]: I1001 10:33:08.467513 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5c7b59dd4f-spxm4_e743cd5e-e8c6-4fe2-9480-3a30316b8e23/proxy-httpd/0.log" Oct 01 10:33:08 crc kubenswrapper[4787]: I1001 10:33:08.696468 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-kkdkg_ed714165-98db-45a6-b96b-69601bff3e56/swift-ring-rebalance/0.log" Oct 01 10:33:08 crc kubenswrapper[4787]: I1001 10:33:08.819638 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3432bc16-99b4-431a-aeb1-600e826bbc3e/account-auditor/0.log" Oct 01 10:33:08 crc kubenswrapper[4787]: I1001 10:33:08.940949 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3432bc16-99b4-431a-aeb1-600e826bbc3e/account-reaper/0.log" Oct 01 10:33:09 crc kubenswrapper[4787]: I1001 10:33:09.074685 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3432bc16-99b4-431a-aeb1-600e826bbc3e/account-replicator/0.log" Oct 01 10:33:09 crc kubenswrapper[4787]: I1001 10:33:09.128615 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3432bc16-99b4-431a-aeb1-600e826bbc3e/account-server/0.log" Oct 01 10:33:09 crc kubenswrapper[4787]: I1001 10:33:09.205334 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3432bc16-99b4-431a-aeb1-600e826bbc3e/container-auditor/0.log" Oct 01 10:33:09 crc kubenswrapper[4787]: I1001 10:33:09.356630 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3432bc16-99b4-431a-aeb1-600e826bbc3e/container-replicator/0.log" Oct 01 10:33:09 crc kubenswrapper[4787]: I1001 10:33:09.374163 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3432bc16-99b4-431a-aeb1-600e826bbc3e/container-server/0.log" Oct 01 10:33:09 crc kubenswrapper[4787]: I1001 10:33:09.506959 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3432bc16-99b4-431a-aeb1-600e826bbc3e/container-updater/0.log" Oct 01 10:33:09 crc kubenswrapper[4787]: I1001 10:33:09.626027 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3432bc16-99b4-431a-aeb1-600e826bbc3e/object-auditor/0.log" Oct 01 10:33:09 crc kubenswrapper[4787]: I1001 10:33:09.695178 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3432bc16-99b4-431a-aeb1-600e826bbc3e/object-expirer/0.log" Oct 01 10:33:09 crc kubenswrapper[4787]: I1001 10:33:09.756285 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3432bc16-99b4-431a-aeb1-600e826bbc3e/object-replicator/0.log" Oct 01 10:33:09 crc kubenswrapper[4787]: I1001 10:33:09.873316 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3432bc16-99b4-431a-aeb1-600e826bbc3e/object-server/0.log" Oct 01 10:33:10 crc kubenswrapper[4787]: I1001 10:33:10.002294 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3432bc16-99b4-431a-aeb1-600e826bbc3e/rsync/0.log" Oct 01 10:33:10 crc kubenswrapper[4787]: I1001 10:33:10.032922 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3432bc16-99b4-431a-aeb1-600e826bbc3e/object-updater/0.log" Oct 01 10:33:10 crc kubenswrapper[4787]: I1001 10:33:10.121463 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3432bc16-99b4-431a-aeb1-600e826bbc3e/swift-recon-cron/0.log" Oct 01 10:33:10 crc kubenswrapper[4787]: I1001 10:33:10.370569 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9_56b1c51b-bdcd-4c55-a00d-105ab4bc04b0/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 10:33:10 crc kubenswrapper[4787]: I1001 10:33:10.556339 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_4b136104-8797-40d5-9ea1-fec8c6be0e90/tempest-tests-tempest-tests-runner/0.log" Oct 01 10:33:10 crc kubenswrapper[4787]: I1001 10:33:10.604826 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_9aaa5922-539c-4e98-99ac-a08a31b14d53/test-operator-logs-container/0.log" Oct 01 10:33:10 crc kubenswrapper[4787]: I1001 10:33:10.846108 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-nkbqd_08369955-6564-47c7-bed0-64893e898e1b/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 10:33:16 crc kubenswrapper[4787]: I1001 10:33:16.650668 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_7137d092-ff92-4b8d-84c5-9d9d269ddcc7/memcached/0.log" Oct 01 10:34:04 crc kubenswrapper[4787]: I1001 10:34:04.588967 4787 generic.go:334] "Generic (PLEG): container finished" podID="1f19a4ae-ba42-42dd-8bd3-1e93aadb2620" containerID="2859f8091b6b21aa4fd604191070c2dd5f6ef835756eb133ca865d2fbda9bce8" exitCode=0 Oct 01 10:34:04 crc kubenswrapper[4787]: I1001 10:34:04.589066 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4twfr/crc-debug-cmf6s" event={"ID":"1f19a4ae-ba42-42dd-8bd3-1e93aadb2620","Type":"ContainerDied","Data":"2859f8091b6b21aa4fd604191070c2dd5f6ef835756eb133ca865d2fbda9bce8"} Oct 01 10:34:05 crc kubenswrapper[4787]: I1001 10:34:05.712794 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4twfr/crc-debug-cmf6s" Oct 01 10:34:05 crc kubenswrapper[4787]: I1001 10:34:05.718789 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqlxd\" (UniqueName: \"kubernetes.io/projected/1f19a4ae-ba42-42dd-8bd3-1e93aadb2620-kube-api-access-cqlxd\") pod \"1f19a4ae-ba42-42dd-8bd3-1e93aadb2620\" (UID: \"1f19a4ae-ba42-42dd-8bd3-1e93aadb2620\") " Oct 01 10:34:05 crc kubenswrapper[4787]: I1001 10:34:05.718905 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1f19a4ae-ba42-42dd-8bd3-1e93aadb2620-host\") pod \"1f19a4ae-ba42-42dd-8bd3-1e93aadb2620\" (UID: \"1f19a4ae-ba42-42dd-8bd3-1e93aadb2620\") " Oct 01 10:34:05 crc kubenswrapper[4787]: I1001 10:34:05.719028 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f19a4ae-ba42-42dd-8bd3-1e93aadb2620-host" (OuterVolumeSpecName: "host") pod "1f19a4ae-ba42-42dd-8bd3-1e93aadb2620" (UID: "1f19a4ae-ba42-42dd-8bd3-1e93aadb2620"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 10:34:05 crc kubenswrapper[4787]: I1001 10:34:05.719576 4787 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1f19a4ae-ba42-42dd-8bd3-1e93aadb2620-host\") on node \"crc\" DevicePath \"\"" Oct 01 10:34:05 crc kubenswrapper[4787]: I1001 10:34:05.735388 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f19a4ae-ba42-42dd-8bd3-1e93aadb2620-kube-api-access-cqlxd" (OuterVolumeSpecName: "kube-api-access-cqlxd") pod "1f19a4ae-ba42-42dd-8bd3-1e93aadb2620" (UID: "1f19a4ae-ba42-42dd-8bd3-1e93aadb2620"). InnerVolumeSpecName "kube-api-access-cqlxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:34:05 crc kubenswrapper[4787]: I1001 10:34:05.745785 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-4twfr/crc-debug-cmf6s"] Oct 01 10:34:05 crc kubenswrapper[4787]: I1001 10:34:05.753898 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-4twfr/crc-debug-cmf6s"] Oct 01 10:34:05 crc kubenswrapper[4787]: I1001 10:34:05.820351 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqlxd\" (UniqueName: \"kubernetes.io/projected/1f19a4ae-ba42-42dd-8bd3-1e93aadb2620-kube-api-access-cqlxd\") on node \"crc\" DevicePath \"\"" Oct 01 10:34:06 crc kubenswrapper[4787]: I1001 10:34:06.544245 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f19a4ae-ba42-42dd-8bd3-1e93aadb2620" path="/var/lib/kubelet/pods/1f19a4ae-ba42-42dd-8bd3-1e93aadb2620/volumes" Oct 01 10:34:06 crc kubenswrapper[4787]: I1001 10:34:06.625331 4787 scope.go:117] "RemoveContainer" containerID="2859f8091b6b21aa4fd604191070c2dd5f6ef835756eb133ca865d2fbda9bce8" Oct 01 10:34:06 crc kubenswrapper[4787]: I1001 10:34:06.625422 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4twfr/crc-debug-cmf6s" Oct 01 10:34:06 crc kubenswrapper[4787]: I1001 10:34:06.966769 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-4twfr/crc-debug-jt92p"] Oct 01 10:34:06 crc kubenswrapper[4787]: E1001 10:34:06.967670 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f19a4ae-ba42-42dd-8bd3-1e93aadb2620" containerName="container-00" Oct 01 10:34:06 crc kubenswrapper[4787]: I1001 10:34:06.967688 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f19a4ae-ba42-42dd-8bd3-1e93aadb2620" containerName="container-00" Oct 01 10:34:06 crc kubenswrapper[4787]: E1001 10:34:06.967701 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e" containerName="extract-utilities" Oct 01 10:34:06 crc kubenswrapper[4787]: I1001 10:34:06.967710 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e" containerName="extract-utilities" Oct 01 10:34:06 crc kubenswrapper[4787]: E1001 10:34:06.967726 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e" containerName="extract-content" Oct 01 10:34:06 crc kubenswrapper[4787]: I1001 10:34:06.967734 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e" containerName="extract-content" Oct 01 10:34:06 crc kubenswrapper[4787]: E1001 10:34:06.967747 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e" containerName="registry-server" Oct 01 10:34:06 crc kubenswrapper[4787]: I1001 10:34:06.967755 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e" containerName="registry-server" Oct 01 10:34:06 crc kubenswrapper[4787]: I1001 10:34:06.967990 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ec43e9a-384c-4cb4-9d49-11a8e3b2a26e" containerName="registry-server" Oct 01 10:34:06 crc kubenswrapper[4787]: I1001 10:34:06.968014 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f19a4ae-ba42-42dd-8bd3-1e93aadb2620" containerName="container-00" Oct 01 10:34:06 crc kubenswrapper[4787]: I1001 10:34:06.968849 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4twfr/crc-debug-jt92p" Oct 01 10:34:06 crc kubenswrapper[4787]: I1001 10:34:06.972501 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-4twfr"/"default-dockercfg-m79wp" Oct 01 10:34:07 crc kubenswrapper[4787]: I1001 10:34:07.158019 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1d13aa0a-9d39-410e-ae67-d9b99804f930-host\") pod \"crc-debug-jt92p\" (UID: \"1d13aa0a-9d39-410e-ae67-d9b99804f930\") " pod="openshift-must-gather-4twfr/crc-debug-jt92p" Oct 01 10:34:07 crc kubenswrapper[4787]: I1001 10:34:07.158182 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwmjz\" (UniqueName: \"kubernetes.io/projected/1d13aa0a-9d39-410e-ae67-d9b99804f930-kube-api-access-vwmjz\") pod \"crc-debug-jt92p\" (UID: \"1d13aa0a-9d39-410e-ae67-d9b99804f930\") " pod="openshift-must-gather-4twfr/crc-debug-jt92p" Oct 01 10:34:07 crc kubenswrapper[4787]: I1001 10:34:07.259968 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwmjz\" (UniqueName: \"kubernetes.io/projected/1d13aa0a-9d39-410e-ae67-d9b99804f930-kube-api-access-vwmjz\") pod \"crc-debug-jt92p\" (UID: \"1d13aa0a-9d39-410e-ae67-d9b99804f930\") " pod="openshift-must-gather-4twfr/crc-debug-jt92p" Oct 01 10:34:07 crc kubenswrapper[4787]: I1001 10:34:07.260236 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1d13aa0a-9d39-410e-ae67-d9b99804f930-host\") pod \"crc-debug-jt92p\" (UID: \"1d13aa0a-9d39-410e-ae67-d9b99804f930\") " pod="openshift-must-gather-4twfr/crc-debug-jt92p" Oct 01 10:34:07 crc kubenswrapper[4787]: I1001 10:34:07.260386 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1d13aa0a-9d39-410e-ae67-d9b99804f930-host\") pod \"crc-debug-jt92p\" (UID: \"1d13aa0a-9d39-410e-ae67-d9b99804f930\") " pod="openshift-must-gather-4twfr/crc-debug-jt92p" Oct 01 10:34:07 crc kubenswrapper[4787]: I1001 10:34:07.279114 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwmjz\" (UniqueName: \"kubernetes.io/projected/1d13aa0a-9d39-410e-ae67-d9b99804f930-kube-api-access-vwmjz\") pod \"crc-debug-jt92p\" (UID: \"1d13aa0a-9d39-410e-ae67-d9b99804f930\") " pod="openshift-must-gather-4twfr/crc-debug-jt92p" Oct 01 10:34:07 crc kubenswrapper[4787]: I1001 10:34:07.288937 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4twfr/crc-debug-jt92p" Oct 01 10:34:07 crc kubenswrapper[4787]: I1001 10:34:07.636539 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4twfr/crc-debug-jt92p" event={"ID":"1d13aa0a-9d39-410e-ae67-d9b99804f930","Type":"ContainerStarted","Data":"dd019f5093eb2e44f07d38ecec1f9ff118048c1e7532ea4677715101811b8bcd"} Oct 01 10:34:07 crc kubenswrapper[4787]: I1001 10:34:07.636997 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4twfr/crc-debug-jt92p" event={"ID":"1d13aa0a-9d39-410e-ae67-d9b99804f930","Type":"ContainerStarted","Data":"87d91ca953266174e369bfffa8df83d66c34bca75dd3a3f1b7681b8d7412cffd"} Oct 01 10:34:07 crc kubenswrapper[4787]: I1001 10:34:07.653097 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-4twfr/crc-debug-jt92p" podStartSLOduration=1.6530413560000001 podStartE2EDuration="1.653041356s" podCreationTimestamp="2025-10-01 10:34:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 10:34:07.648809443 +0000 UTC m=+3479.763953620" watchObservedRunningTime="2025-10-01 10:34:07.653041356 +0000 UTC m=+3479.768185513" Oct 01 10:34:08 crc kubenswrapper[4787]: I1001 10:34:08.651740 4787 generic.go:334] "Generic (PLEG): container finished" podID="1d13aa0a-9d39-410e-ae67-d9b99804f930" containerID="dd019f5093eb2e44f07d38ecec1f9ff118048c1e7532ea4677715101811b8bcd" exitCode=0 Oct 01 10:34:08 crc kubenswrapper[4787]: I1001 10:34:08.651865 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4twfr/crc-debug-jt92p" event={"ID":"1d13aa0a-9d39-410e-ae67-d9b99804f930","Type":"ContainerDied","Data":"dd019f5093eb2e44f07d38ecec1f9ff118048c1e7532ea4677715101811b8bcd"} Oct 01 10:34:09 crc kubenswrapper[4787]: I1001 10:34:09.778469 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4twfr/crc-debug-jt92p" Oct 01 10:34:09 crc kubenswrapper[4787]: I1001 10:34:09.896877 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwmjz\" (UniqueName: \"kubernetes.io/projected/1d13aa0a-9d39-410e-ae67-d9b99804f930-kube-api-access-vwmjz\") pod \"1d13aa0a-9d39-410e-ae67-d9b99804f930\" (UID: \"1d13aa0a-9d39-410e-ae67-d9b99804f930\") " Oct 01 10:34:09 crc kubenswrapper[4787]: I1001 10:34:09.897175 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1d13aa0a-9d39-410e-ae67-d9b99804f930-host\") pod \"1d13aa0a-9d39-410e-ae67-d9b99804f930\" (UID: \"1d13aa0a-9d39-410e-ae67-d9b99804f930\") " Oct 01 10:34:09 crc kubenswrapper[4787]: I1001 10:34:09.897216 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1d13aa0a-9d39-410e-ae67-d9b99804f930-host" (OuterVolumeSpecName: "host") pod "1d13aa0a-9d39-410e-ae67-d9b99804f930" (UID: "1d13aa0a-9d39-410e-ae67-d9b99804f930"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 10:34:09 crc kubenswrapper[4787]: I1001 10:34:09.897571 4787 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1d13aa0a-9d39-410e-ae67-d9b99804f930-host\") on node \"crc\" DevicePath \"\"" Oct 01 10:34:09 crc kubenswrapper[4787]: I1001 10:34:09.902635 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d13aa0a-9d39-410e-ae67-d9b99804f930-kube-api-access-vwmjz" (OuterVolumeSpecName: "kube-api-access-vwmjz") pod "1d13aa0a-9d39-410e-ae67-d9b99804f930" (UID: "1d13aa0a-9d39-410e-ae67-d9b99804f930"). InnerVolumeSpecName "kube-api-access-vwmjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:34:09 crc kubenswrapper[4787]: I1001 10:34:09.998842 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwmjz\" (UniqueName: \"kubernetes.io/projected/1d13aa0a-9d39-410e-ae67-d9b99804f930-kube-api-access-vwmjz\") on node \"crc\" DevicePath \"\"" Oct 01 10:34:10 crc kubenswrapper[4787]: I1001 10:34:10.684595 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4twfr/crc-debug-jt92p" event={"ID":"1d13aa0a-9d39-410e-ae67-d9b99804f930","Type":"ContainerDied","Data":"87d91ca953266174e369bfffa8df83d66c34bca75dd3a3f1b7681b8d7412cffd"} Oct 01 10:34:10 crc kubenswrapper[4787]: I1001 10:34:10.684920 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="87d91ca953266174e369bfffa8df83d66c34bca75dd3a3f1b7681b8d7412cffd" Oct 01 10:34:10 crc kubenswrapper[4787]: I1001 10:34:10.684637 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4twfr/crc-debug-jt92p" Oct 01 10:34:11 crc kubenswrapper[4787]: I1001 10:34:11.250264 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:34:11 crc kubenswrapper[4787]: I1001 10:34:11.250344 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:34:14 crc kubenswrapper[4787]: I1001 10:34:14.422060 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-4twfr/crc-debug-jt92p"] Oct 01 10:34:14 crc kubenswrapper[4787]: I1001 10:34:14.432646 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-4twfr/crc-debug-jt92p"] Oct 01 10:34:14 crc kubenswrapper[4787]: I1001 10:34:14.534302 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d13aa0a-9d39-410e-ae67-d9b99804f930" path="/var/lib/kubelet/pods/1d13aa0a-9d39-410e-ae67-d9b99804f930/volumes" Oct 01 10:34:15 crc kubenswrapper[4787]: I1001 10:34:15.572737 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-4twfr/crc-debug-fxr59"] Oct 01 10:34:15 crc kubenswrapper[4787]: E1001 10:34:15.574554 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d13aa0a-9d39-410e-ae67-d9b99804f930" containerName="container-00" Oct 01 10:34:15 crc kubenswrapper[4787]: I1001 10:34:15.574697 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d13aa0a-9d39-410e-ae67-d9b99804f930" containerName="container-00" Oct 01 10:34:15 crc kubenswrapper[4787]: I1001 10:34:15.575061 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d13aa0a-9d39-410e-ae67-d9b99804f930" containerName="container-00" Oct 01 10:34:15 crc kubenswrapper[4787]: I1001 10:34:15.576034 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4twfr/crc-debug-fxr59" Oct 01 10:34:15 crc kubenswrapper[4787]: I1001 10:34:15.578494 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-4twfr"/"default-dockercfg-m79wp" Oct 01 10:34:15 crc kubenswrapper[4787]: I1001 10:34:15.586328 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9m8g5\" (UniqueName: \"kubernetes.io/projected/e663d69a-20aa-4b37-a2ae-18691e94bbaf-kube-api-access-9m8g5\") pod \"crc-debug-fxr59\" (UID: \"e663d69a-20aa-4b37-a2ae-18691e94bbaf\") " pod="openshift-must-gather-4twfr/crc-debug-fxr59" Oct 01 10:34:15 crc kubenswrapper[4787]: I1001 10:34:15.586407 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e663d69a-20aa-4b37-a2ae-18691e94bbaf-host\") pod \"crc-debug-fxr59\" (UID: \"e663d69a-20aa-4b37-a2ae-18691e94bbaf\") " pod="openshift-must-gather-4twfr/crc-debug-fxr59" Oct 01 10:34:15 crc kubenswrapper[4787]: I1001 10:34:15.689221 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9m8g5\" (UniqueName: \"kubernetes.io/projected/e663d69a-20aa-4b37-a2ae-18691e94bbaf-kube-api-access-9m8g5\") pod \"crc-debug-fxr59\" (UID: \"e663d69a-20aa-4b37-a2ae-18691e94bbaf\") " pod="openshift-must-gather-4twfr/crc-debug-fxr59" Oct 01 10:34:15 crc kubenswrapper[4787]: I1001 10:34:15.689737 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e663d69a-20aa-4b37-a2ae-18691e94bbaf-host\") pod \"crc-debug-fxr59\" (UID: \"e663d69a-20aa-4b37-a2ae-18691e94bbaf\") " pod="openshift-must-gather-4twfr/crc-debug-fxr59" Oct 01 10:34:15 crc kubenswrapper[4787]: I1001 10:34:15.689809 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e663d69a-20aa-4b37-a2ae-18691e94bbaf-host\") pod \"crc-debug-fxr59\" (UID: \"e663d69a-20aa-4b37-a2ae-18691e94bbaf\") " pod="openshift-must-gather-4twfr/crc-debug-fxr59" Oct 01 10:34:15 crc kubenswrapper[4787]: I1001 10:34:15.709507 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9m8g5\" (UniqueName: \"kubernetes.io/projected/e663d69a-20aa-4b37-a2ae-18691e94bbaf-kube-api-access-9m8g5\") pod \"crc-debug-fxr59\" (UID: \"e663d69a-20aa-4b37-a2ae-18691e94bbaf\") " pod="openshift-must-gather-4twfr/crc-debug-fxr59" Oct 01 10:34:15 crc kubenswrapper[4787]: I1001 10:34:15.895606 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4twfr/crc-debug-fxr59" Oct 01 10:34:16 crc kubenswrapper[4787]: I1001 10:34:16.740551 4787 generic.go:334] "Generic (PLEG): container finished" podID="e663d69a-20aa-4b37-a2ae-18691e94bbaf" containerID="71cb4c32e7098030731770915a830f0105f024ca5dd937ae9c5ea4c8f290ec7f" exitCode=0 Oct 01 10:34:16 crc kubenswrapper[4787]: I1001 10:34:16.740605 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4twfr/crc-debug-fxr59" event={"ID":"e663d69a-20aa-4b37-a2ae-18691e94bbaf","Type":"ContainerDied","Data":"71cb4c32e7098030731770915a830f0105f024ca5dd937ae9c5ea4c8f290ec7f"} Oct 01 10:34:16 crc kubenswrapper[4787]: I1001 10:34:16.740636 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4twfr/crc-debug-fxr59" event={"ID":"e663d69a-20aa-4b37-a2ae-18691e94bbaf","Type":"ContainerStarted","Data":"b278401974ac93adbcfd0427f892a1bec2a01639d0fe8e4acf1c071ed7fbaf96"} Oct 01 10:34:16 crc kubenswrapper[4787]: I1001 10:34:16.782452 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-4twfr/crc-debug-fxr59"] Oct 01 10:34:16 crc kubenswrapper[4787]: I1001 10:34:16.789870 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-4twfr/crc-debug-fxr59"] Oct 01 10:34:17 crc kubenswrapper[4787]: I1001 10:34:17.855661 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4twfr/crc-debug-fxr59" Oct 01 10:34:17 crc kubenswrapper[4787]: I1001 10:34:17.947646 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9m8g5\" (UniqueName: \"kubernetes.io/projected/e663d69a-20aa-4b37-a2ae-18691e94bbaf-kube-api-access-9m8g5\") pod \"e663d69a-20aa-4b37-a2ae-18691e94bbaf\" (UID: \"e663d69a-20aa-4b37-a2ae-18691e94bbaf\") " Oct 01 10:34:17 crc kubenswrapper[4787]: I1001 10:34:17.947829 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e663d69a-20aa-4b37-a2ae-18691e94bbaf-host\") pod \"e663d69a-20aa-4b37-a2ae-18691e94bbaf\" (UID: \"e663d69a-20aa-4b37-a2ae-18691e94bbaf\") " Oct 01 10:34:17 crc kubenswrapper[4787]: I1001 10:34:17.948337 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e663d69a-20aa-4b37-a2ae-18691e94bbaf-host" (OuterVolumeSpecName: "host") pod "e663d69a-20aa-4b37-a2ae-18691e94bbaf" (UID: "e663d69a-20aa-4b37-a2ae-18691e94bbaf"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 10:34:17 crc kubenswrapper[4787]: I1001 10:34:17.955190 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e663d69a-20aa-4b37-a2ae-18691e94bbaf-kube-api-access-9m8g5" (OuterVolumeSpecName: "kube-api-access-9m8g5") pod "e663d69a-20aa-4b37-a2ae-18691e94bbaf" (UID: "e663d69a-20aa-4b37-a2ae-18691e94bbaf"). InnerVolumeSpecName "kube-api-access-9m8g5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:34:18 crc kubenswrapper[4787]: I1001 10:34:18.049546 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9m8g5\" (UniqueName: \"kubernetes.io/projected/e663d69a-20aa-4b37-a2ae-18691e94bbaf-kube-api-access-9m8g5\") on node \"crc\" DevicePath \"\"" Oct 01 10:34:18 crc kubenswrapper[4787]: I1001 10:34:18.049579 4787 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e663d69a-20aa-4b37-a2ae-18691e94bbaf-host\") on node \"crc\" DevicePath \"\"" Oct 01 10:34:18 crc kubenswrapper[4787]: I1001 10:34:18.485248 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh_a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1/util/0.log" Oct 01 10:34:18 crc kubenswrapper[4787]: I1001 10:34:18.535380 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e663d69a-20aa-4b37-a2ae-18691e94bbaf" path="/var/lib/kubelet/pods/e663d69a-20aa-4b37-a2ae-18691e94bbaf/volumes" Oct 01 10:34:18 crc kubenswrapper[4787]: I1001 10:34:18.693985 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh_a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1/util/0.log" Oct 01 10:34:18 crc kubenswrapper[4787]: I1001 10:34:18.746810 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh_a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1/pull/0.log" Oct 01 10:34:18 crc kubenswrapper[4787]: I1001 10:34:18.747169 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh_a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1/pull/0.log" Oct 01 10:34:18 crc kubenswrapper[4787]: I1001 10:34:18.761693 4787 scope.go:117] "RemoveContainer" containerID="71cb4c32e7098030731770915a830f0105f024ca5dd937ae9c5ea4c8f290ec7f" Oct 01 10:34:18 crc kubenswrapper[4787]: I1001 10:34:18.761721 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4twfr/crc-debug-fxr59" Oct 01 10:34:18 crc kubenswrapper[4787]: I1001 10:34:18.902538 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh_a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1/util/0.log" Oct 01 10:34:18 crc kubenswrapper[4787]: I1001 10:34:18.944620 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh_a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1/pull/0.log" Oct 01 10:34:18 crc kubenswrapper[4787]: I1001 10:34:18.991604 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh_a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1/extract/0.log" Oct 01 10:34:19 crc kubenswrapper[4787]: I1001 10:34:19.110546 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-f7f98cb69-wqs6n_26b292ae-dcb2-41f6-86b5-2eddfecdf356/kube-rbac-proxy/0.log" Oct 01 10:34:19 crc kubenswrapper[4787]: I1001 10:34:19.243689 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859cd486d-6h8cm_24354b04-0daa-4fd4-8d38-b21f64bc2cfe/kube-rbac-proxy/0.log" Oct 01 10:34:19 crc kubenswrapper[4787]: I1001 10:34:19.275470 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-f7f98cb69-wqs6n_26b292ae-dcb2-41f6-86b5-2eddfecdf356/manager/0.log" Oct 01 10:34:19 crc kubenswrapper[4787]: I1001 10:34:19.424377 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859cd486d-6h8cm_24354b04-0daa-4fd4-8d38-b21f64bc2cfe/manager/0.log" Oct 01 10:34:19 crc kubenswrapper[4787]: I1001 10:34:19.437051 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-77fb7bcf5b-2pxzs_77e07ae2-1dae-4570-84b8-c137e86ac628/kube-rbac-proxy/0.log" Oct 01 10:34:19 crc kubenswrapper[4787]: I1001 10:34:19.475200 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-77fb7bcf5b-2pxzs_77e07ae2-1dae-4570-84b8-c137e86ac628/manager/0.log" Oct 01 10:34:19 crc kubenswrapper[4787]: I1001 10:34:19.643765 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-8bc4775b5-x2v89_b4902328-fb90-40f9-a84c-5a0359d699d4/kube-rbac-proxy/0.log" Oct 01 10:34:19 crc kubenswrapper[4787]: I1001 10:34:19.851819 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-8bc4775b5-x2v89_b4902328-fb90-40f9-a84c-5a0359d699d4/manager/0.log" Oct 01 10:34:19 crc kubenswrapper[4787]: I1001 10:34:19.890840 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b4fc86755-lbbxd_b1a0ae18-3c6d-473a-a660-5ff59f6a87bf/kube-rbac-proxy/0.log" Oct 01 10:34:19 crc kubenswrapper[4787]: I1001 10:34:19.927407 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b4fc86755-lbbxd_b1a0ae18-3c6d-473a-a660-5ff59f6a87bf/manager/0.log" Oct 01 10:34:20 crc kubenswrapper[4787]: I1001 10:34:20.045777 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-679b4759bb-qt9bq_8f52292a-2603-4524-b3fc-8d7b003f95f8/kube-rbac-proxy/0.log" Oct 01 10:34:20 crc kubenswrapper[4787]: I1001 10:34:20.123548 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-679b4759bb-qt9bq_8f52292a-2603-4524-b3fc-8d7b003f95f8/manager/0.log" Oct 01 10:34:20 crc kubenswrapper[4787]: I1001 10:34:20.179766 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5c8fdc4d5c-t72zp_9e03d54f-d462-4894-9c3b-8253a861af54/kube-rbac-proxy/0.log" Oct 01 10:34:20 crc kubenswrapper[4787]: I1001 10:34:20.370216 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5f45cd594f-zrtv2_ae2ba2f0-8422-4a03-b64f-d335f76c5532/kube-rbac-proxy/0.log" Oct 01 10:34:20 crc kubenswrapper[4787]: I1001 10:34:20.432870 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5c8fdc4d5c-t72zp_9e03d54f-d462-4894-9c3b-8253a861af54/manager/0.log" Oct 01 10:34:20 crc kubenswrapper[4787]: I1001 10:34:20.482578 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5f45cd594f-zrtv2_ae2ba2f0-8422-4a03-b64f-d335f76c5532/manager/0.log" Oct 01 10:34:20 crc kubenswrapper[4787]: I1001 10:34:20.665471 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-59d7dc95cf-9tdp2_88d93cbc-27c2-47be-a882-6607925d4391/kube-rbac-proxy/0.log" Oct 01 10:34:20 crc kubenswrapper[4787]: I1001 10:34:20.746832 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-59d7dc95cf-9tdp2_88d93cbc-27c2-47be-a882-6607925d4391/manager/0.log" Oct 01 10:34:20 crc kubenswrapper[4787]: I1001 10:34:20.919374 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-b7cf8cb5f-5qz2m_2dfc2636-198f-4b51-b07e-40ff76c1091c/manager/0.log" Oct 01 10:34:20 crc kubenswrapper[4787]: I1001 10:34:20.958396 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-b7cf8cb5f-5qz2m_2dfc2636-198f-4b51-b07e-40ff76c1091c/kube-rbac-proxy/0.log" Oct 01 10:34:21 crc kubenswrapper[4787]: I1001 10:34:21.017041 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-67bf5bb885-qmr4v_b54a691a-cd96-4ae4-a9ce-65662f58cd60/kube-rbac-proxy/0.log" Oct 01 10:34:21 crc kubenswrapper[4787]: I1001 10:34:21.141188 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-67bf5bb885-qmr4v_b54a691a-cd96-4ae4-a9ce-65662f58cd60/manager/0.log" Oct 01 10:34:21 crc kubenswrapper[4787]: I1001 10:34:21.223123 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-54fbbfcd44-4jlcz_8b59a5e1-4f45-4bcd-b9ce-63bae8b5ed57/kube-rbac-proxy/0.log" Oct 01 10:34:21 crc kubenswrapper[4787]: I1001 10:34:21.350305 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-54fbbfcd44-4jlcz_8b59a5e1-4f45-4bcd-b9ce-63bae8b5ed57/manager/0.log" Oct 01 10:34:21 crc kubenswrapper[4787]: I1001 10:34:21.472565 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7fd5b6bbc6-nbxwx_c186a6f1-61b9-4ab7-8996-5e1d8439a77c/kube-rbac-proxy/0.log" Oct 01 10:34:21 crc kubenswrapper[4787]: I1001 10:34:21.544465 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7fd5b6bbc6-nbxwx_c186a6f1-61b9-4ab7-8996-5e1d8439a77c/manager/0.log" Oct 01 10:34:21 crc kubenswrapper[4787]: I1001 10:34:21.597843 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-75f8d67d86-xjbwt_20e0b9a6-f51b-4925-bffd-8fa00e9a8ec7/kube-rbac-proxy/0.log" Oct 01 10:34:21 crc kubenswrapper[4787]: I1001 10:34:21.665491 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-75f8d67d86-xjbwt_20e0b9a6-f51b-4925-bffd-8fa00e9a8ec7/manager/0.log" Oct 01 10:34:21 crc kubenswrapper[4787]: I1001 10:34:21.787958 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-659bb84579xswrt_2d5ffd79-67fa-4dd0-81a7-b44156946b86/manager/0.log" Oct 01 10:34:21 crc kubenswrapper[4787]: I1001 10:34:21.814835 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-659bb84579xswrt_2d5ffd79-67fa-4dd0-81a7-b44156946b86/kube-rbac-proxy/0.log" Oct 01 10:34:21 crc kubenswrapper[4787]: I1001 10:34:21.947049 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6c7b6bcb7c-bxglr_40d23171-497b-47d9-96f4-51f32d7683a1/kube-rbac-proxy/0.log" Oct 01 10:34:22 crc kubenswrapper[4787]: I1001 10:34:22.046153 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-676c66f88b-7hgdc_faaa86ca-d16f-4ee1-9a9a-15c54aca2a8b/kube-rbac-proxy/0.log" Oct 01 10:34:22 crc kubenswrapper[4787]: I1001 10:34:22.402912 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-676c66f88b-7hgdc_faaa86ca-d16f-4ee1-9a9a-15c54aca2a8b/operator/0.log" Oct 01 10:34:22 crc kubenswrapper[4787]: I1001 10:34:22.426788 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-jnbw4_c6b4eeb7-99c2-44ff-9f71-e27c17c17649/registry-server/0.log" Oct 01 10:34:22 crc kubenswrapper[4787]: I1001 10:34:22.486047 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-84c745747f-8qtcw_57f9bf14-0074-43ba-882d-a9102b047a3b/kube-rbac-proxy/0.log" Oct 01 10:34:22 crc kubenswrapper[4787]: I1001 10:34:22.607221 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-84c745747f-8qtcw_57f9bf14-0074-43ba-882d-a9102b047a3b/manager/0.log" Oct 01 10:34:22 crc kubenswrapper[4787]: I1001 10:34:22.643418 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-598c4c8547-4zm2q_5024aa9b-39fc-4e58-96fe-c66d054e3d5a/kube-rbac-proxy/0.log" Oct 01 10:34:22 crc kubenswrapper[4787]: I1001 10:34:22.777840 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-598c4c8547-4zm2q_5024aa9b-39fc-4e58-96fe-c66d054e3d5a/manager/0.log" Oct 01 10:34:22 crc kubenswrapper[4787]: I1001 10:34:22.914776 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-xvd4r_f31cb2c2-eabc-4bca-88c5-2aa194fb13b5/operator/0.log" Oct 01 10:34:23 crc kubenswrapper[4787]: I1001 10:34:23.039003 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-689b4f76c9-xvrsn_ed13cb34-e846-41ab-aff7-8e5d4acd2e6b/kube-rbac-proxy/0.log" Oct 01 10:34:23 crc kubenswrapper[4787]: I1001 10:34:23.185408 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-689b4f76c9-xvrsn_ed13cb34-e846-41ab-aff7-8e5d4acd2e6b/manager/0.log" Oct 01 10:34:23 crc kubenswrapper[4787]: I1001 10:34:23.295709 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6c7b6bcb7c-bxglr_40d23171-497b-47d9-96f4-51f32d7683a1/manager/0.log" Oct 01 10:34:23 crc kubenswrapper[4787]: I1001 10:34:23.299835 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-cb66d6b59-fqgnk_7139ac70-d405-449b-924e-01fd94e1375a/kube-rbac-proxy/0.log" Oct 01 10:34:23 crc kubenswrapper[4787]: I1001 10:34:23.322002 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-cb66d6b59-fqgnk_7139ac70-d405-449b-924e-01fd94e1375a/manager/0.log" Oct 01 10:34:23 crc kubenswrapper[4787]: I1001 10:34:23.422039 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-cbdf6dc66-z2gzf_a74d851b-a805-4138-9bf4-1a4546a2f30a/kube-rbac-proxy/0.log" Oct 01 10:34:23 crc kubenswrapper[4787]: I1001 10:34:23.517099 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-cbdf6dc66-z2gzf_a74d851b-a805-4138-9bf4-1a4546a2f30a/manager/0.log" Oct 01 10:34:23 crc kubenswrapper[4787]: I1001 10:34:23.529415 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-68d7bc5569-4mgdw_fb776e43-5f95-44f0-96c0-945b2681d8c1/kube-rbac-proxy/0.log" Oct 01 10:34:23 crc kubenswrapper[4787]: I1001 10:34:23.653755 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-68d7bc5569-4mgdw_fb776e43-5f95-44f0-96c0-945b2681d8c1/manager/0.log" Oct 01 10:34:38 crc kubenswrapper[4787]: I1001 10:34:38.720298 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-zssrz_e6b988c0-936d-4b49-80a2-015ae390f262/control-plane-machine-set-operator/0.log" Oct 01 10:34:38 crc kubenswrapper[4787]: I1001 10:34:38.912762 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-4g5xc_07892f22-556b-49a6-8f71-3d3b16b2cdef/kube-rbac-proxy/0.log" Oct 01 10:34:38 crc kubenswrapper[4787]: I1001 10:34:38.922930 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-4g5xc_07892f22-556b-49a6-8f71-3d3b16b2cdef/machine-api-operator/0.log" Oct 01 10:34:41 crc kubenswrapper[4787]: I1001 10:34:41.250398 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:34:41 crc kubenswrapper[4787]: I1001 10:34:41.250718 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:34:48 crc kubenswrapper[4787]: I1001 10:34:48.190754 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-472tp"] Oct 01 10:34:48 crc kubenswrapper[4787]: E1001 10:34:48.195322 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e663d69a-20aa-4b37-a2ae-18691e94bbaf" containerName="container-00" Oct 01 10:34:48 crc kubenswrapper[4787]: I1001 10:34:48.195352 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e663d69a-20aa-4b37-a2ae-18691e94bbaf" containerName="container-00" Oct 01 10:34:48 crc kubenswrapper[4787]: I1001 10:34:48.195857 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="e663d69a-20aa-4b37-a2ae-18691e94bbaf" containerName="container-00" Oct 01 10:34:48 crc kubenswrapper[4787]: I1001 10:34:48.205461 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-472tp" Oct 01 10:34:48 crc kubenswrapper[4787]: I1001 10:34:48.210887 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-472tp"] Oct 01 10:34:48 crc kubenswrapper[4787]: I1001 10:34:48.324258 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1-catalog-content\") pod \"certified-operators-472tp\" (UID: \"5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1\") " pod="openshift-marketplace/certified-operators-472tp" Oct 01 10:34:48 crc kubenswrapper[4787]: I1001 10:34:48.324618 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1-utilities\") pod \"certified-operators-472tp\" (UID: \"5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1\") " pod="openshift-marketplace/certified-operators-472tp" Oct 01 10:34:48 crc kubenswrapper[4787]: I1001 10:34:48.324705 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km5hp\" (UniqueName: \"kubernetes.io/projected/5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1-kube-api-access-km5hp\") pod \"certified-operators-472tp\" (UID: \"5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1\") " pod="openshift-marketplace/certified-operators-472tp" Oct 01 10:34:48 crc kubenswrapper[4787]: I1001 10:34:48.426490 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1-utilities\") pod \"certified-operators-472tp\" (UID: \"5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1\") " pod="openshift-marketplace/certified-operators-472tp" Oct 01 10:34:48 crc kubenswrapper[4787]: I1001 10:34:48.426656 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km5hp\" (UniqueName: \"kubernetes.io/projected/5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1-kube-api-access-km5hp\") pod \"certified-operators-472tp\" (UID: \"5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1\") " pod="openshift-marketplace/certified-operators-472tp" Oct 01 10:34:48 crc kubenswrapper[4787]: I1001 10:34:48.426742 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1-catalog-content\") pod \"certified-operators-472tp\" (UID: \"5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1\") " pod="openshift-marketplace/certified-operators-472tp" Oct 01 10:34:48 crc kubenswrapper[4787]: I1001 10:34:48.426989 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1-utilities\") pod \"certified-operators-472tp\" (UID: \"5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1\") " pod="openshift-marketplace/certified-operators-472tp" Oct 01 10:34:48 crc kubenswrapper[4787]: I1001 10:34:48.427026 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1-catalog-content\") pod \"certified-operators-472tp\" (UID: \"5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1\") " pod="openshift-marketplace/certified-operators-472tp" Oct 01 10:34:48 crc kubenswrapper[4787]: I1001 10:34:48.450611 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km5hp\" (UniqueName: \"kubernetes.io/projected/5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1-kube-api-access-km5hp\") pod \"certified-operators-472tp\" (UID: \"5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1\") " pod="openshift-marketplace/certified-operators-472tp" Oct 01 10:34:48 crc kubenswrapper[4787]: I1001 10:34:48.526548 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-472tp" Oct 01 10:34:49 crc kubenswrapper[4787]: I1001 10:34:49.019375 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-472tp"] Oct 01 10:34:49 crc kubenswrapper[4787]: I1001 10:34:49.108498 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-472tp" event={"ID":"5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1","Type":"ContainerStarted","Data":"1ccda87ebd96fc1daaecba94b479182f47740f9515b4512f422b1226b41e2c73"} Oct 01 10:34:50 crc kubenswrapper[4787]: I1001 10:34:50.101369 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-w2pvt_b1f30c7b-06ed-4ccf-a3d0-8df65a3a8c67/cert-manager-controller/0.log" Oct 01 10:34:50 crc kubenswrapper[4787]: I1001 10:34:50.134988 4787 generic.go:334] "Generic (PLEG): container finished" podID="5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1" containerID="20ad44dec4c8154432c68f3db5600cbea862a678ee211cb32776e07c2abc8291" exitCode=0 Oct 01 10:34:50 crc kubenswrapper[4787]: I1001 10:34:50.135038 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-472tp" event={"ID":"5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1","Type":"ContainerDied","Data":"20ad44dec4c8154432c68f3db5600cbea862a678ee211cb32776e07c2abc8291"} Oct 01 10:34:50 crc kubenswrapper[4787]: I1001 10:34:50.290744 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-ss8nf_081d3606-6bc1-4f9b-8f3b-6ee00ed9d164/cert-manager-cainjector/0.log" Oct 01 10:34:50 crc kubenswrapper[4787]: I1001 10:34:50.296607 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-xgrnp_426ff9a0-bcf0-486e-bd80-2d4de51be0cf/cert-manager-webhook/0.log" Oct 01 10:34:51 crc kubenswrapper[4787]: I1001 10:34:51.147364 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-472tp" event={"ID":"5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1","Type":"ContainerStarted","Data":"1ab896a6c6be5cd11e34ff4b40ecda1018b3d3b33dc71ec8e6c4a9128001878b"} Oct 01 10:34:52 crc kubenswrapper[4787]: I1001 10:34:52.158223 4787 generic.go:334] "Generic (PLEG): container finished" podID="5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1" containerID="1ab896a6c6be5cd11e34ff4b40ecda1018b3d3b33dc71ec8e6c4a9128001878b" exitCode=0 Oct 01 10:34:52 crc kubenswrapper[4787]: I1001 10:34:52.158305 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-472tp" event={"ID":"5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1","Type":"ContainerDied","Data":"1ab896a6c6be5cd11e34ff4b40ecda1018b3d3b33dc71ec8e6c4a9128001878b"} Oct 01 10:34:53 crc kubenswrapper[4787]: I1001 10:34:53.181052 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-472tp" event={"ID":"5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1","Type":"ContainerStarted","Data":"f4d878a261d35146f16409c80ffe87387fec67f07ed9f262cb5ced80af53ce2b"} Oct 01 10:34:53 crc kubenswrapper[4787]: I1001 10:34:53.204749 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-472tp" podStartSLOduration=2.668814815 podStartE2EDuration="5.204729643s" podCreationTimestamp="2025-10-01 10:34:48 +0000 UTC" firstStartedPulling="2025-10-01 10:34:50.138031986 +0000 UTC m=+3522.253176143" lastFinishedPulling="2025-10-01 10:34:52.673946814 +0000 UTC m=+3524.789090971" observedRunningTime="2025-10-01 10:34:53.199510545 +0000 UTC m=+3525.314654712" watchObservedRunningTime="2025-10-01 10:34:53.204729643 +0000 UTC m=+3525.319873800" Oct 01 10:34:58 crc kubenswrapper[4787]: I1001 10:34:58.536022 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-472tp" Oct 01 10:34:58 crc kubenswrapper[4787]: I1001 10:34:58.536438 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-472tp" Oct 01 10:34:58 crc kubenswrapper[4787]: I1001 10:34:58.578133 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-472tp" Oct 01 10:34:59 crc kubenswrapper[4787]: I1001 10:34:59.279528 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-472tp" Oct 01 10:34:59 crc kubenswrapper[4787]: I1001 10:34:59.349469 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-472tp"] Oct 01 10:35:01 crc kubenswrapper[4787]: I1001 10:35:01.246504 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-472tp" podUID="5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1" containerName="registry-server" containerID="cri-o://f4d878a261d35146f16409c80ffe87387fec67f07ed9f262cb5ced80af53ce2b" gracePeriod=2 Oct 01 10:35:01 crc kubenswrapper[4787]: I1001 10:35:01.364626 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-864bb6dfb5-cjhqm_e4a805b1-d488-4007-9dd3-38d3fbab1f10/nmstate-console-plugin/0.log" Oct 01 10:35:01 crc kubenswrapper[4787]: I1001 10:35:01.583745 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-2rzr2_ee9e051d-7391-4a43-9a02-5084da577b36/nmstate-handler/0.log" Oct 01 10:35:01 crc kubenswrapper[4787]: I1001 10:35:01.666183 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58fcddf996-2cnst_10cd7a64-357c-4c0d-b11c-426e4060b61c/nmstate-metrics/0.log" Oct 01 10:35:01 crc kubenswrapper[4787]: I1001 10:35:01.754363 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58fcddf996-2cnst_10cd7a64-357c-4c0d-b11c-426e4060b61c/kube-rbac-proxy/0.log" Oct 01 10:35:01 crc kubenswrapper[4787]: I1001 10:35:01.775162 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-472tp" Oct 01 10:35:01 crc kubenswrapper[4787]: I1001 10:35:01.887175 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5d6f6cfd66-8l5nr_27ab4c98-35a2-45be-864e-4550be0fa875/nmstate-operator/0.log" Oct 01 10:35:01 crc kubenswrapper[4787]: I1001 10:35:01.897131 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-km5hp\" (UniqueName: \"kubernetes.io/projected/5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1-kube-api-access-km5hp\") pod \"5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1\" (UID: \"5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1\") " Oct 01 10:35:01 crc kubenswrapper[4787]: I1001 10:35:01.897291 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1-catalog-content\") pod \"5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1\" (UID: \"5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1\") " Oct 01 10:35:01 crc kubenswrapper[4787]: I1001 10:35:01.897426 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1-utilities\") pod \"5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1\" (UID: \"5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1\") " Oct 01 10:35:01 crc kubenswrapper[4787]: I1001 10:35:01.898471 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1-utilities" (OuterVolumeSpecName: "utilities") pod "5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1" (UID: "5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:35:01 crc kubenswrapper[4787]: I1001 10:35:01.916284 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1-kube-api-access-km5hp" (OuterVolumeSpecName: "kube-api-access-km5hp") pod "5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1" (UID: "5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1"). InnerVolumeSpecName "kube-api-access-km5hp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:35:02 crc kubenswrapper[4787]: I1001 10:35:02.000283 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-km5hp\" (UniqueName: \"kubernetes.io/projected/5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1-kube-api-access-km5hp\") on node \"crc\" DevicePath \"\"" Oct 01 10:35:02 crc kubenswrapper[4787]: I1001 10:35:02.000325 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 10:35:02 crc kubenswrapper[4787]: I1001 10:35:02.042286 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6d689559c5-22sp6_97efd9a6-e221-4aac-97b1-b49de45edd88/nmstate-webhook/0.log" Oct 01 10:35:02 crc kubenswrapper[4787]: I1001 10:35:02.224844 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1" (UID: "5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:35:02 crc kubenswrapper[4787]: I1001 10:35:02.257916 4787 generic.go:334] "Generic (PLEG): container finished" podID="5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1" containerID="f4d878a261d35146f16409c80ffe87387fec67f07ed9f262cb5ced80af53ce2b" exitCode=0 Oct 01 10:35:02 crc kubenswrapper[4787]: I1001 10:35:02.257964 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-472tp" event={"ID":"5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1","Type":"ContainerDied","Data":"f4d878a261d35146f16409c80ffe87387fec67f07ed9f262cb5ced80af53ce2b"} Oct 01 10:35:02 crc kubenswrapper[4787]: I1001 10:35:02.257994 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-472tp" event={"ID":"5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1","Type":"ContainerDied","Data":"1ccda87ebd96fc1daaecba94b479182f47740f9515b4512f422b1226b41e2c73"} Oct 01 10:35:02 crc kubenswrapper[4787]: I1001 10:35:02.258015 4787 scope.go:117] "RemoveContainer" containerID="f4d878a261d35146f16409c80ffe87387fec67f07ed9f262cb5ced80af53ce2b" Oct 01 10:35:02 crc kubenswrapper[4787]: I1001 10:35:02.258049 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-472tp" Oct 01 10:35:02 crc kubenswrapper[4787]: I1001 10:35:02.280583 4787 scope.go:117] "RemoveContainer" containerID="1ab896a6c6be5cd11e34ff4b40ecda1018b3d3b33dc71ec8e6c4a9128001878b" Oct 01 10:35:02 crc kubenswrapper[4787]: I1001 10:35:02.295113 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-472tp"] Oct 01 10:35:02 crc kubenswrapper[4787]: I1001 10:35:02.304433 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-472tp"] Oct 01 10:35:02 crc kubenswrapper[4787]: I1001 10:35:02.305031 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 10:35:02 crc kubenswrapper[4787]: I1001 10:35:02.323274 4787 scope.go:117] "RemoveContainer" containerID="20ad44dec4c8154432c68f3db5600cbea862a678ee211cb32776e07c2abc8291" Oct 01 10:35:02 crc kubenswrapper[4787]: I1001 10:35:02.346041 4787 scope.go:117] "RemoveContainer" containerID="f4d878a261d35146f16409c80ffe87387fec67f07ed9f262cb5ced80af53ce2b" Oct 01 10:35:02 crc kubenswrapper[4787]: E1001 10:35:02.346704 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4d878a261d35146f16409c80ffe87387fec67f07ed9f262cb5ced80af53ce2b\": container with ID starting with f4d878a261d35146f16409c80ffe87387fec67f07ed9f262cb5ced80af53ce2b not found: ID does not exist" containerID="f4d878a261d35146f16409c80ffe87387fec67f07ed9f262cb5ced80af53ce2b" Oct 01 10:35:02 crc kubenswrapper[4787]: I1001 10:35:02.346736 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4d878a261d35146f16409c80ffe87387fec67f07ed9f262cb5ced80af53ce2b"} err="failed to get container status \"f4d878a261d35146f16409c80ffe87387fec67f07ed9f262cb5ced80af53ce2b\": rpc error: code = NotFound desc = could not find container \"f4d878a261d35146f16409c80ffe87387fec67f07ed9f262cb5ced80af53ce2b\": container with ID starting with f4d878a261d35146f16409c80ffe87387fec67f07ed9f262cb5ced80af53ce2b not found: ID does not exist" Oct 01 10:35:02 crc kubenswrapper[4787]: I1001 10:35:02.346760 4787 scope.go:117] "RemoveContainer" containerID="1ab896a6c6be5cd11e34ff4b40ecda1018b3d3b33dc71ec8e6c4a9128001878b" Oct 01 10:35:02 crc kubenswrapper[4787]: E1001 10:35:02.347158 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ab896a6c6be5cd11e34ff4b40ecda1018b3d3b33dc71ec8e6c4a9128001878b\": container with ID starting with 1ab896a6c6be5cd11e34ff4b40ecda1018b3d3b33dc71ec8e6c4a9128001878b not found: ID does not exist" containerID="1ab896a6c6be5cd11e34ff4b40ecda1018b3d3b33dc71ec8e6c4a9128001878b" Oct 01 10:35:02 crc kubenswrapper[4787]: I1001 10:35:02.347208 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ab896a6c6be5cd11e34ff4b40ecda1018b3d3b33dc71ec8e6c4a9128001878b"} err="failed to get container status \"1ab896a6c6be5cd11e34ff4b40ecda1018b3d3b33dc71ec8e6c4a9128001878b\": rpc error: code = NotFound desc = could not find container \"1ab896a6c6be5cd11e34ff4b40ecda1018b3d3b33dc71ec8e6c4a9128001878b\": container with ID starting with 1ab896a6c6be5cd11e34ff4b40ecda1018b3d3b33dc71ec8e6c4a9128001878b not found: ID does not exist" Oct 01 10:35:02 crc kubenswrapper[4787]: I1001 10:35:02.347240 4787 scope.go:117] "RemoveContainer" containerID="20ad44dec4c8154432c68f3db5600cbea862a678ee211cb32776e07c2abc8291" Oct 01 10:35:02 crc kubenswrapper[4787]: E1001 10:35:02.347592 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20ad44dec4c8154432c68f3db5600cbea862a678ee211cb32776e07c2abc8291\": container with ID starting with 20ad44dec4c8154432c68f3db5600cbea862a678ee211cb32776e07c2abc8291 not found: ID does not exist" containerID="20ad44dec4c8154432c68f3db5600cbea862a678ee211cb32776e07c2abc8291" Oct 01 10:35:02 crc kubenswrapper[4787]: I1001 10:35:02.347619 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20ad44dec4c8154432c68f3db5600cbea862a678ee211cb32776e07c2abc8291"} err="failed to get container status \"20ad44dec4c8154432c68f3db5600cbea862a678ee211cb32776e07c2abc8291\": rpc error: code = NotFound desc = could not find container \"20ad44dec4c8154432c68f3db5600cbea862a678ee211cb32776e07c2abc8291\": container with ID starting with 20ad44dec4c8154432c68f3db5600cbea862a678ee211cb32776e07c2abc8291 not found: ID does not exist" Oct 01 10:35:02 crc kubenswrapper[4787]: I1001 10:35:02.536116 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1" path="/var/lib/kubelet/pods/5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1/volumes" Oct 01 10:35:11 crc kubenswrapper[4787]: I1001 10:35:11.250779 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:35:11 crc kubenswrapper[4787]: I1001 10:35:11.251394 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:35:11 crc kubenswrapper[4787]: I1001 10:35:11.251443 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" Oct 01 10:35:11 crc kubenswrapper[4787]: I1001 10:35:11.252182 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7cb2890ef30bbede4935afca82be38385def845a6f5b1c0bd5934ac785342b56"} pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 10:35:11 crc kubenswrapper[4787]: I1001 10:35:11.252246 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" containerID="cri-o://7cb2890ef30bbede4935afca82be38385def845a6f5b1c0bd5934ac785342b56" gracePeriod=600 Oct 01 10:35:11 crc kubenswrapper[4787]: E1001 10:35:11.377401 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:35:12 crc kubenswrapper[4787]: I1001 10:35:12.351122 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" event={"ID":"fa51b883-5f3e-4141-9d4c-37704aac7718","Type":"ContainerDied","Data":"7cb2890ef30bbede4935afca82be38385def845a6f5b1c0bd5934ac785342b56"} Oct 01 10:35:12 crc kubenswrapper[4787]: I1001 10:35:12.351114 4787 generic.go:334] "Generic (PLEG): container finished" podID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerID="7cb2890ef30bbede4935afca82be38385def845a6f5b1c0bd5934ac785342b56" exitCode=0 Oct 01 10:35:12 crc kubenswrapper[4787]: I1001 10:35:12.351517 4787 scope.go:117] "RemoveContainer" containerID="600ca70fa1dc39a1299d76f2c60704fbc1ec67b77ac9063a4f3e5239fa4265de" Oct 01 10:35:12 crc kubenswrapper[4787]: I1001 10:35:12.352230 4787 scope.go:117] "RemoveContainer" containerID="7cb2890ef30bbede4935afca82be38385def845a6f5b1c0bd5934ac785342b56" Oct 01 10:35:12 crc kubenswrapper[4787]: E1001 10:35:12.352590 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:35:15 crc kubenswrapper[4787]: I1001 10:35:15.166316 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5d688f5ffc-gnclz_9a498324-bbec-40e1-9649-f23ca26561f3/kube-rbac-proxy/0.log" Oct 01 10:35:15 crc kubenswrapper[4787]: I1001 10:35:15.285066 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5d688f5ffc-gnclz_9a498324-bbec-40e1-9649-f23ca26561f3/controller/0.log" Oct 01 10:35:15 crc kubenswrapper[4787]: I1001 10:35:15.403541 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-csxhg_c6570c99-4cff-4194-9744-0efc00f88332/cp-frr-files/0.log" Oct 01 10:35:15 crc kubenswrapper[4787]: I1001 10:35:15.493900 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-csxhg_c6570c99-4cff-4194-9744-0efc00f88332/cp-frr-files/0.log" Oct 01 10:35:15 crc kubenswrapper[4787]: I1001 10:35:15.513467 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-csxhg_c6570c99-4cff-4194-9744-0efc00f88332/cp-metrics/0.log" Oct 01 10:35:15 crc kubenswrapper[4787]: I1001 10:35:15.523378 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-csxhg_c6570c99-4cff-4194-9744-0efc00f88332/cp-reloader/0.log" Oct 01 10:35:15 crc kubenswrapper[4787]: I1001 10:35:15.612439 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-csxhg_c6570c99-4cff-4194-9744-0efc00f88332/cp-reloader/0.log" Oct 01 10:35:15 crc kubenswrapper[4787]: I1001 10:35:15.780882 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-csxhg_c6570c99-4cff-4194-9744-0efc00f88332/cp-metrics/0.log" Oct 01 10:35:15 crc kubenswrapper[4787]: I1001 10:35:15.787980 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-csxhg_c6570c99-4cff-4194-9744-0efc00f88332/cp-metrics/0.log" Oct 01 10:35:15 crc kubenswrapper[4787]: I1001 10:35:15.796823 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-csxhg_c6570c99-4cff-4194-9744-0efc00f88332/cp-reloader/0.log" Oct 01 10:35:15 crc kubenswrapper[4787]: I1001 10:35:15.847602 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-csxhg_c6570c99-4cff-4194-9744-0efc00f88332/cp-frr-files/0.log" Oct 01 10:35:16 crc kubenswrapper[4787]: I1001 10:35:16.025471 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-csxhg_c6570c99-4cff-4194-9744-0efc00f88332/controller/0.log" Oct 01 10:35:16 crc kubenswrapper[4787]: I1001 10:35:16.028569 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-csxhg_c6570c99-4cff-4194-9744-0efc00f88332/cp-reloader/0.log" Oct 01 10:35:16 crc kubenswrapper[4787]: I1001 10:35:16.028584 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-csxhg_c6570c99-4cff-4194-9744-0efc00f88332/cp-frr-files/0.log" Oct 01 10:35:16 crc kubenswrapper[4787]: I1001 10:35:16.040739 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-csxhg_c6570c99-4cff-4194-9744-0efc00f88332/cp-metrics/0.log" Oct 01 10:35:16 crc kubenswrapper[4787]: I1001 10:35:16.241901 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-csxhg_c6570c99-4cff-4194-9744-0efc00f88332/kube-rbac-proxy/0.log" Oct 01 10:35:16 crc kubenswrapper[4787]: I1001 10:35:16.257151 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-csxhg_c6570c99-4cff-4194-9744-0efc00f88332/kube-rbac-proxy-frr/0.log" Oct 01 10:35:16 crc kubenswrapper[4787]: I1001 10:35:16.265930 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-csxhg_c6570c99-4cff-4194-9744-0efc00f88332/frr-metrics/0.log" Oct 01 10:35:16 crc kubenswrapper[4787]: I1001 10:35:16.431663 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-csxhg_c6570c99-4cff-4194-9744-0efc00f88332/reloader/0.log" Oct 01 10:35:16 crc kubenswrapper[4787]: I1001 10:35:16.475584 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-5478bdb765-kcxb8_8b191fd2-3066-4e07-9641-98e44be362f5/frr-k8s-webhook-server/0.log" Oct 01 10:35:16 crc kubenswrapper[4787]: I1001 10:35:16.726550 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-86f7dcd4cf-c4xjr_178c8464-8b87-4dc1-9c52-7dc1f5b3714d/manager/0.log" Oct 01 10:35:16 crc kubenswrapper[4787]: I1001 10:35:16.831918 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-7d48896987-v5r98_55c631b4-c7e0-47cb-9e66-93b8a8c51877/webhook-server/0.log" Oct 01 10:35:16 crc kubenswrapper[4787]: I1001 10:35:16.958282 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-cc45k_c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668/kube-rbac-proxy/0.log" Oct 01 10:35:17 crc kubenswrapper[4787]: I1001 10:35:17.536248 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-cc45k_c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668/speaker/0.log" Oct 01 10:35:17 crc kubenswrapper[4787]: I1001 10:35:17.696019 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-csxhg_c6570c99-4cff-4194-9744-0efc00f88332/frr/0.log" Oct 01 10:35:26 crc kubenswrapper[4787]: I1001 10:35:26.524830 4787 scope.go:117] "RemoveContainer" containerID="7cb2890ef30bbede4935afca82be38385def845a6f5b1c0bd5934ac785342b56" Oct 01 10:35:26 crc kubenswrapper[4787]: E1001 10:35:26.525683 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:35:29 crc kubenswrapper[4787]: I1001 10:35:29.331299 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc_e60e8301-91e9-45d5-a3d1-cd29ce1e9514/util/0.log" Oct 01 10:35:29 crc kubenswrapper[4787]: I1001 10:35:29.617931 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc_e60e8301-91e9-45d5-a3d1-cd29ce1e9514/pull/0.log" Oct 01 10:35:29 crc kubenswrapper[4787]: I1001 10:35:29.624131 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc_e60e8301-91e9-45d5-a3d1-cd29ce1e9514/util/0.log" Oct 01 10:35:29 crc kubenswrapper[4787]: I1001 10:35:29.673823 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc_e60e8301-91e9-45d5-a3d1-cd29ce1e9514/pull/0.log" Oct 01 10:35:30 crc kubenswrapper[4787]: I1001 10:35:30.037615 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc_e60e8301-91e9-45d5-a3d1-cd29ce1e9514/pull/0.log" Oct 01 10:35:30 crc kubenswrapper[4787]: I1001 10:35:30.051088 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc_e60e8301-91e9-45d5-a3d1-cd29ce1e9514/util/0.log" Oct 01 10:35:30 crc kubenswrapper[4787]: I1001 10:35:30.057770 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc_e60e8301-91e9-45d5-a3d1-cd29ce1e9514/extract/0.log" Oct 01 10:35:30 crc kubenswrapper[4787]: I1001 10:35:30.263329 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8ksjp_09d257f5-ca13-42c9-92f8-920f99ebadc1/extract-utilities/0.log" Oct 01 10:35:30 crc kubenswrapper[4787]: I1001 10:35:30.461611 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8ksjp_09d257f5-ca13-42c9-92f8-920f99ebadc1/extract-content/0.log" Oct 01 10:35:30 crc kubenswrapper[4787]: I1001 10:35:30.472664 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8ksjp_09d257f5-ca13-42c9-92f8-920f99ebadc1/extract-utilities/0.log" Oct 01 10:35:30 crc kubenswrapper[4787]: I1001 10:35:30.498968 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8ksjp_09d257f5-ca13-42c9-92f8-920f99ebadc1/extract-content/0.log" Oct 01 10:35:30 crc kubenswrapper[4787]: I1001 10:35:30.682524 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8ksjp_09d257f5-ca13-42c9-92f8-920f99ebadc1/extract-content/0.log" Oct 01 10:35:30 crc kubenswrapper[4787]: I1001 10:35:30.728572 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8ksjp_09d257f5-ca13-42c9-92f8-920f99ebadc1/extract-utilities/0.log" Oct 01 10:35:30 crc kubenswrapper[4787]: I1001 10:35:30.957256 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4m4pm_6e4be47d-87d5-469d-9ad2-b12efb75e1aa/extract-utilities/0.log" Oct 01 10:35:31 crc kubenswrapper[4787]: I1001 10:35:31.161600 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4m4pm_6e4be47d-87d5-469d-9ad2-b12efb75e1aa/extract-utilities/0.log" Oct 01 10:35:31 crc kubenswrapper[4787]: I1001 10:35:31.266617 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4m4pm_6e4be47d-87d5-469d-9ad2-b12efb75e1aa/extract-content/0.log" Oct 01 10:35:31 crc kubenswrapper[4787]: I1001 10:35:31.277114 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8ksjp_09d257f5-ca13-42c9-92f8-920f99ebadc1/registry-server/0.log" Oct 01 10:35:31 crc kubenswrapper[4787]: I1001 10:35:31.289801 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4m4pm_6e4be47d-87d5-469d-9ad2-b12efb75e1aa/extract-content/0.log" Oct 01 10:35:31 crc kubenswrapper[4787]: I1001 10:35:31.467593 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4m4pm_6e4be47d-87d5-469d-9ad2-b12efb75e1aa/extract-utilities/0.log" Oct 01 10:35:31 crc kubenswrapper[4787]: I1001 10:35:31.481543 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4m4pm_6e4be47d-87d5-469d-9ad2-b12efb75e1aa/extract-content/0.log" Oct 01 10:35:31 crc kubenswrapper[4787]: I1001 10:35:31.768559 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft_810c9efa-ff37-4f0a-8c30-d620a00bd8a4/util/0.log" Oct 01 10:35:31 crc kubenswrapper[4787]: I1001 10:35:31.915820 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft_810c9efa-ff37-4f0a-8c30-d620a00bd8a4/pull/0.log" Oct 01 10:35:31 crc kubenswrapper[4787]: I1001 10:35:31.998915 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft_810c9efa-ff37-4f0a-8c30-d620a00bd8a4/util/0.log" Oct 01 10:35:32 crc kubenswrapper[4787]: I1001 10:35:32.083539 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft_810c9efa-ff37-4f0a-8c30-d620a00bd8a4/pull/0.log" Oct 01 10:35:32 crc kubenswrapper[4787]: I1001 10:35:32.171693 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4m4pm_6e4be47d-87d5-469d-9ad2-b12efb75e1aa/registry-server/0.log" Oct 01 10:35:32 crc kubenswrapper[4787]: I1001 10:35:32.203518 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft_810c9efa-ff37-4f0a-8c30-d620a00bd8a4/util/0.log" Oct 01 10:35:32 crc kubenswrapper[4787]: I1001 10:35:32.244054 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft_810c9efa-ff37-4f0a-8c30-d620a00bd8a4/pull/0.log" Oct 01 10:35:32 crc kubenswrapper[4787]: I1001 10:35:32.314017 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft_810c9efa-ff37-4f0a-8c30-d620a00bd8a4/extract/0.log" Oct 01 10:35:32 crc kubenswrapper[4787]: I1001 10:35:32.427528 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-6zb6m_3390abc0-6474-4a34-8e95-c5923f699e80/marketplace-operator/0.log" Oct 01 10:35:32 crc kubenswrapper[4787]: I1001 10:35:32.574224 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-swbhm_593f4070-9419-4e2f-9da0-741e86877588/extract-utilities/0.log" Oct 01 10:35:32 crc kubenswrapper[4787]: I1001 10:35:32.709182 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-swbhm_593f4070-9419-4e2f-9da0-741e86877588/extract-utilities/0.log" Oct 01 10:35:32 crc kubenswrapper[4787]: I1001 10:35:32.717508 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-swbhm_593f4070-9419-4e2f-9da0-741e86877588/extract-content/0.log" Oct 01 10:35:32 crc kubenswrapper[4787]: I1001 10:35:32.731216 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-swbhm_593f4070-9419-4e2f-9da0-741e86877588/extract-content/0.log" Oct 01 10:35:32 crc kubenswrapper[4787]: I1001 10:35:32.968504 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-swbhm_593f4070-9419-4e2f-9da0-741e86877588/extract-content/0.log" Oct 01 10:35:33 crc kubenswrapper[4787]: I1001 10:35:33.004716 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-swbhm_593f4070-9419-4e2f-9da0-741e86877588/extract-utilities/0.log" Oct 01 10:35:33 crc kubenswrapper[4787]: I1001 10:35:33.136265 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-swbhm_593f4070-9419-4e2f-9da0-741e86877588/registry-server/0.log" Oct 01 10:35:33 crc kubenswrapper[4787]: I1001 10:35:33.152335 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pwj4z_c3868e91-d922-4768-af6a-123795df5a28/extract-utilities/0.log" Oct 01 10:35:33 crc kubenswrapper[4787]: I1001 10:35:33.352013 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pwj4z_c3868e91-d922-4768-af6a-123795df5a28/extract-content/0.log" Oct 01 10:35:33 crc kubenswrapper[4787]: I1001 10:35:33.389931 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pwj4z_c3868e91-d922-4768-af6a-123795df5a28/extract-content/0.log" Oct 01 10:35:33 crc kubenswrapper[4787]: I1001 10:35:33.407053 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pwj4z_c3868e91-d922-4768-af6a-123795df5a28/extract-utilities/0.log" Oct 01 10:35:33 crc kubenswrapper[4787]: I1001 10:35:33.547559 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pwj4z_c3868e91-d922-4768-af6a-123795df5a28/extract-utilities/0.log" Oct 01 10:35:33 crc kubenswrapper[4787]: I1001 10:35:33.553810 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pwj4z_c3868e91-d922-4768-af6a-123795df5a28/extract-content/0.log" Oct 01 10:35:33 crc kubenswrapper[4787]: I1001 10:35:33.980094 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pwj4z_c3868e91-d922-4768-af6a-123795df5a28/registry-server/0.log" Oct 01 10:35:40 crc kubenswrapper[4787]: I1001 10:35:40.523821 4787 scope.go:117] "RemoveContainer" containerID="7cb2890ef30bbede4935afca82be38385def845a6f5b1c0bd5934ac785342b56" Oct 01 10:35:40 crc kubenswrapper[4787]: E1001 10:35:40.524600 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:35:55 crc kubenswrapper[4787]: I1001 10:35:55.524690 4787 scope.go:117] "RemoveContainer" containerID="7cb2890ef30bbede4935afca82be38385def845a6f5b1c0bd5934ac785342b56" Oct 01 10:35:55 crc kubenswrapper[4787]: E1001 10:35:55.525456 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:36:07 crc kubenswrapper[4787]: I1001 10:36:07.523506 4787 scope.go:117] "RemoveContainer" containerID="7cb2890ef30bbede4935afca82be38385def845a6f5b1c0bd5934ac785342b56" Oct 01 10:36:07 crc kubenswrapper[4787]: E1001 10:36:07.524283 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:36:19 crc kubenswrapper[4787]: I1001 10:36:19.524630 4787 scope.go:117] "RemoveContainer" containerID="7cb2890ef30bbede4935afca82be38385def845a6f5b1c0bd5934ac785342b56" Oct 01 10:36:19 crc kubenswrapper[4787]: E1001 10:36:19.525488 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:36:32 crc kubenswrapper[4787]: I1001 10:36:32.524023 4787 scope.go:117] "RemoveContainer" containerID="7cb2890ef30bbede4935afca82be38385def845a6f5b1c0bd5934ac785342b56" Oct 01 10:36:32 crc kubenswrapper[4787]: E1001 10:36:32.524915 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:36:47 crc kubenswrapper[4787]: I1001 10:36:47.524421 4787 scope.go:117] "RemoveContainer" containerID="7cb2890ef30bbede4935afca82be38385def845a6f5b1c0bd5934ac785342b56" Oct 01 10:36:47 crc kubenswrapper[4787]: E1001 10:36:47.525321 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:36:58 crc kubenswrapper[4787]: I1001 10:36:58.533316 4787 scope.go:117] "RemoveContainer" containerID="7cb2890ef30bbede4935afca82be38385def845a6f5b1c0bd5934ac785342b56" Oct 01 10:36:58 crc kubenswrapper[4787]: E1001 10:36:58.535180 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:37:13 crc kubenswrapper[4787]: I1001 10:37:13.523930 4787 scope.go:117] "RemoveContainer" containerID="7cb2890ef30bbede4935afca82be38385def845a6f5b1c0bd5934ac785342b56" Oct 01 10:37:13 crc kubenswrapper[4787]: E1001 10:37:13.524757 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:37:25 crc kubenswrapper[4787]: I1001 10:37:25.523859 4787 scope.go:117] "RemoveContainer" containerID="7cb2890ef30bbede4935afca82be38385def845a6f5b1c0bd5934ac785342b56" Oct 01 10:37:25 crc kubenswrapper[4787]: E1001 10:37:25.524629 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:37:38 crc kubenswrapper[4787]: I1001 10:37:38.534543 4787 scope.go:117] "RemoveContainer" containerID="7cb2890ef30bbede4935afca82be38385def845a6f5b1c0bd5934ac785342b56" Oct 01 10:37:38 crc kubenswrapper[4787]: E1001 10:37:38.535906 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:37:40 crc kubenswrapper[4787]: I1001 10:37:40.792438 4787 generic.go:334] "Generic (PLEG): container finished" podID="cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe" containerID="b4f7cff751b06a5d5c173da30c66c55eb2db181c7d012a6ab5963c9665355dd4" exitCode=0 Oct 01 10:37:40 crc kubenswrapper[4787]: I1001 10:37:40.792556 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4twfr/must-gather-89b86" event={"ID":"cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe","Type":"ContainerDied","Data":"b4f7cff751b06a5d5c173da30c66c55eb2db181c7d012a6ab5963c9665355dd4"} Oct 01 10:37:40 crc kubenswrapper[4787]: I1001 10:37:40.793818 4787 scope.go:117] "RemoveContainer" containerID="b4f7cff751b06a5d5c173da30c66c55eb2db181c7d012a6ab5963c9665355dd4" Oct 01 10:37:41 crc kubenswrapper[4787]: I1001 10:37:41.273232 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-4twfr_must-gather-89b86_cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe/gather/0.log" Oct 01 10:37:49 crc kubenswrapper[4787]: I1001 10:37:49.954863 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-4twfr/must-gather-89b86"] Oct 01 10:37:49 crc kubenswrapper[4787]: I1001 10:37:49.955818 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-4twfr/must-gather-89b86" podUID="cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe" containerName="copy" containerID="cri-o://4aa83081c78d868deba718968a1a161f88c8c2c16bca9fcf3ca5ad6a06dc8700" gracePeriod=2 Oct 01 10:37:49 crc kubenswrapper[4787]: I1001 10:37:49.967906 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-4twfr/must-gather-89b86"] Oct 01 10:37:50 crc kubenswrapper[4787]: I1001 10:37:50.524171 4787 scope.go:117] "RemoveContainer" containerID="7cb2890ef30bbede4935afca82be38385def845a6f5b1c0bd5934ac785342b56" Oct 01 10:37:50 crc kubenswrapper[4787]: E1001 10:37:50.525041 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:37:50 crc kubenswrapper[4787]: I1001 10:37:50.596258 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-4twfr_must-gather-89b86_cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe/copy/0.log" Oct 01 10:37:50 crc kubenswrapper[4787]: I1001 10:37:50.596614 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4twfr/must-gather-89b86" Oct 01 10:37:50 crc kubenswrapper[4787]: I1001 10:37:50.710537 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe-must-gather-output\") pod \"cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe\" (UID: \"cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe\") " Oct 01 10:37:50 crc kubenswrapper[4787]: I1001 10:37:50.710773 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grprh\" (UniqueName: \"kubernetes.io/projected/cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe-kube-api-access-grprh\") pod \"cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe\" (UID: \"cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe\") " Oct 01 10:37:50 crc kubenswrapper[4787]: I1001 10:37:50.721397 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe-kube-api-access-grprh" (OuterVolumeSpecName: "kube-api-access-grprh") pod "cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe" (UID: "cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe"). InnerVolumeSpecName "kube-api-access-grprh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:37:50 crc kubenswrapper[4787]: I1001 10:37:50.813028 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grprh\" (UniqueName: \"kubernetes.io/projected/cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe-kube-api-access-grprh\") on node \"crc\" DevicePath \"\"" Oct 01 10:37:50 crc kubenswrapper[4787]: I1001 10:37:50.873155 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe" (UID: "cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:37:50 crc kubenswrapper[4787]: I1001 10:37:50.905907 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-4twfr_must-gather-89b86_cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe/copy/0.log" Oct 01 10:37:50 crc kubenswrapper[4787]: I1001 10:37:50.906692 4787 generic.go:334] "Generic (PLEG): container finished" podID="cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe" containerID="4aa83081c78d868deba718968a1a161f88c8c2c16bca9fcf3ca5ad6a06dc8700" exitCode=143 Oct 01 10:37:50 crc kubenswrapper[4787]: I1001 10:37:50.906820 4787 scope.go:117] "RemoveContainer" containerID="4aa83081c78d868deba718968a1a161f88c8c2c16bca9fcf3ca5ad6a06dc8700" Oct 01 10:37:50 crc kubenswrapper[4787]: I1001 10:37:50.906898 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4twfr/must-gather-89b86" Oct 01 10:37:50 crc kubenswrapper[4787]: I1001 10:37:50.914658 4787 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 01 10:37:50 crc kubenswrapper[4787]: I1001 10:37:50.932790 4787 scope.go:117] "RemoveContainer" containerID="b4f7cff751b06a5d5c173da30c66c55eb2db181c7d012a6ab5963c9665355dd4" Oct 01 10:37:50 crc kubenswrapper[4787]: I1001 10:37:50.989716 4787 scope.go:117] "RemoveContainer" containerID="4aa83081c78d868deba718968a1a161f88c8c2c16bca9fcf3ca5ad6a06dc8700" Oct 01 10:37:50 crc kubenswrapper[4787]: E1001 10:37:50.990161 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4aa83081c78d868deba718968a1a161f88c8c2c16bca9fcf3ca5ad6a06dc8700\": container with ID starting with 4aa83081c78d868deba718968a1a161f88c8c2c16bca9fcf3ca5ad6a06dc8700 not found: ID does not exist" containerID="4aa83081c78d868deba718968a1a161f88c8c2c16bca9fcf3ca5ad6a06dc8700" Oct 01 10:37:50 crc kubenswrapper[4787]: I1001 10:37:50.990209 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4aa83081c78d868deba718968a1a161f88c8c2c16bca9fcf3ca5ad6a06dc8700"} err="failed to get container status \"4aa83081c78d868deba718968a1a161f88c8c2c16bca9fcf3ca5ad6a06dc8700\": rpc error: code = NotFound desc = could not find container \"4aa83081c78d868deba718968a1a161f88c8c2c16bca9fcf3ca5ad6a06dc8700\": container with ID starting with 4aa83081c78d868deba718968a1a161f88c8c2c16bca9fcf3ca5ad6a06dc8700 not found: ID does not exist" Oct 01 10:37:50 crc kubenswrapper[4787]: I1001 10:37:50.990237 4787 scope.go:117] "RemoveContainer" containerID="b4f7cff751b06a5d5c173da30c66c55eb2db181c7d012a6ab5963c9665355dd4" Oct 01 10:37:50 crc kubenswrapper[4787]: E1001 10:37:50.990555 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4f7cff751b06a5d5c173da30c66c55eb2db181c7d012a6ab5963c9665355dd4\": container with ID starting with b4f7cff751b06a5d5c173da30c66c55eb2db181c7d012a6ab5963c9665355dd4 not found: ID does not exist" containerID="b4f7cff751b06a5d5c173da30c66c55eb2db181c7d012a6ab5963c9665355dd4" Oct 01 10:37:50 crc kubenswrapper[4787]: I1001 10:37:50.990601 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4f7cff751b06a5d5c173da30c66c55eb2db181c7d012a6ab5963c9665355dd4"} err="failed to get container status \"b4f7cff751b06a5d5c173da30c66c55eb2db181c7d012a6ab5963c9665355dd4\": rpc error: code = NotFound desc = could not find container \"b4f7cff751b06a5d5c173da30c66c55eb2db181c7d012a6ab5963c9665355dd4\": container with ID starting with b4f7cff751b06a5d5c173da30c66c55eb2db181c7d012a6ab5963c9665355dd4 not found: ID does not exist" Oct 01 10:37:52 crc kubenswrapper[4787]: I1001 10:37:52.534812 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe" path="/var/lib/kubelet/pods/cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe/volumes" Oct 01 10:38:02 crc kubenswrapper[4787]: I1001 10:38:02.524051 4787 scope.go:117] "RemoveContainer" containerID="7cb2890ef30bbede4935afca82be38385def845a6f5b1c0bd5934ac785342b56" Oct 01 10:38:02 crc kubenswrapper[4787]: E1001 10:38:02.527826 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:38:15 crc kubenswrapper[4787]: I1001 10:38:15.524993 4787 scope.go:117] "RemoveContainer" containerID="7cb2890ef30bbede4935afca82be38385def845a6f5b1c0bd5934ac785342b56" Oct 01 10:38:15 crc kubenswrapper[4787]: E1001 10:38:15.526718 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:38:29 crc kubenswrapper[4787]: I1001 10:38:29.523919 4787 scope.go:117] "RemoveContainer" containerID="7cb2890ef30bbede4935afca82be38385def845a6f5b1c0bd5934ac785342b56" Oct 01 10:38:29 crc kubenswrapper[4787]: E1001 10:38:29.525418 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:38:36 crc kubenswrapper[4787]: I1001 10:38:36.317148 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-x2bwg/must-gather-ht4gk"] Oct 01 10:38:36 crc kubenswrapper[4787]: E1001 10:38:36.318340 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1" containerName="extract-utilities" Oct 01 10:38:36 crc kubenswrapper[4787]: I1001 10:38:36.318354 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1" containerName="extract-utilities" Oct 01 10:38:36 crc kubenswrapper[4787]: E1001 10:38:36.318368 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1" containerName="registry-server" Oct 01 10:38:36 crc kubenswrapper[4787]: I1001 10:38:36.318374 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1" containerName="registry-server" Oct 01 10:38:36 crc kubenswrapper[4787]: E1001 10:38:36.318396 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe" containerName="copy" Oct 01 10:38:36 crc kubenswrapper[4787]: I1001 10:38:36.318402 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe" containerName="copy" Oct 01 10:38:36 crc kubenswrapper[4787]: E1001 10:38:36.318418 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1" containerName="extract-content" Oct 01 10:38:36 crc kubenswrapper[4787]: I1001 10:38:36.318425 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1" containerName="extract-content" Oct 01 10:38:36 crc kubenswrapper[4787]: E1001 10:38:36.318437 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe" containerName="gather" Oct 01 10:38:36 crc kubenswrapper[4787]: I1001 10:38:36.318443 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe" containerName="gather" Oct 01 10:38:36 crc kubenswrapper[4787]: I1001 10:38:36.318619 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c37fcd6-4e9f-4cb2-b3ab-3eece8a1a7c1" containerName="registry-server" Oct 01 10:38:36 crc kubenswrapper[4787]: I1001 10:38:36.318641 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe" containerName="gather" Oct 01 10:38:36 crc kubenswrapper[4787]: I1001 10:38:36.318653 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf3c25d6-1bdf-4bb6-b85d-cf31aea04dbe" containerName="copy" Oct 01 10:38:36 crc kubenswrapper[4787]: I1001 10:38:36.326220 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x2bwg/must-gather-ht4gk" Oct 01 10:38:36 crc kubenswrapper[4787]: I1001 10:38:36.341378 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-x2bwg"/"openshift-service-ca.crt" Oct 01 10:38:36 crc kubenswrapper[4787]: I1001 10:38:36.341897 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-x2bwg"/"kube-root-ca.crt" Oct 01 10:38:36 crc kubenswrapper[4787]: I1001 10:38:36.351476 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-x2bwg/must-gather-ht4gk"] Oct 01 10:38:36 crc kubenswrapper[4787]: I1001 10:38:36.465715 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/32eecbc7-b51a-465f-84ce-5352da7cf52a-must-gather-output\") pod \"must-gather-ht4gk\" (UID: \"32eecbc7-b51a-465f-84ce-5352da7cf52a\") " pod="openshift-must-gather-x2bwg/must-gather-ht4gk" Oct 01 10:38:36 crc kubenswrapper[4787]: I1001 10:38:36.465877 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ft755\" (UniqueName: \"kubernetes.io/projected/32eecbc7-b51a-465f-84ce-5352da7cf52a-kube-api-access-ft755\") pod \"must-gather-ht4gk\" (UID: \"32eecbc7-b51a-465f-84ce-5352da7cf52a\") " pod="openshift-must-gather-x2bwg/must-gather-ht4gk" Oct 01 10:38:36 crc kubenswrapper[4787]: I1001 10:38:36.568826 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/32eecbc7-b51a-465f-84ce-5352da7cf52a-must-gather-output\") pod \"must-gather-ht4gk\" (UID: \"32eecbc7-b51a-465f-84ce-5352da7cf52a\") " pod="openshift-must-gather-x2bwg/must-gather-ht4gk" Oct 01 10:38:36 crc kubenswrapper[4787]: I1001 10:38:36.569031 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ft755\" (UniqueName: \"kubernetes.io/projected/32eecbc7-b51a-465f-84ce-5352da7cf52a-kube-api-access-ft755\") pod \"must-gather-ht4gk\" (UID: \"32eecbc7-b51a-465f-84ce-5352da7cf52a\") " pod="openshift-must-gather-x2bwg/must-gather-ht4gk" Oct 01 10:38:36 crc kubenswrapper[4787]: I1001 10:38:36.569385 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/32eecbc7-b51a-465f-84ce-5352da7cf52a-must-gather-output\") pod \"must-gather-ht4gk\" (UID: \"32eecbc7-b51a-465f-84ce-5352da7cf52a\") " pod="openshift-must-gather-x2bwg/must-gather-ht4gk" Oct 01 10:38:36 crc kubenswrapper[4787]: I1001 10:38:36.590163 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ft755\" (UniqueName: \"kubernetes.io/projected/32eecbc7-b51a-465f-84ce-5352da7cf52a-kube-api-access-ft755\") pod \"must-gather-ht4gk\" (UID: \"32eecbc7-b51a-465f-84ce-5352da7cf52a\") " pod="openshift-must-gather-x2bwg/must-gather-ht4gk" Oct 01 10:38:36 crc kubenswrapper[4787]: I1001 10:38:36.653740 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x2bwg/must-gather-ht4gk" Oct 01 10:38:37 crc kubenswrapper[4787]: I1001 10:38:37.970128 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-x2bwg/must-gather-ht4gk"] Oct 01 10:38:38 crc kubenswrapper[4787]: I1001 10:38:38.356372 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x2bwg/must-gather-ht4gk" event={"ID":"32eecbc7-b51a-465f-84ce-5352da7cf52a","Type":"ContainerStarted","Data":"704ff570cc23c833b0c60e0dcab0dc4cf9e99374f16310c9733fea7a04de15f9"} Oct 01 10:38:39 crc kubenswrapper[4787]: I1001 10:38:39.402600 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x2bwg/must-gather-ht4gk" event={"ID":"32eecbc7-b51a-465f-84ce-5352da7cf52a","Type":"ContainerStarted","Data":"7922e468653e12e1e904dbd354bc84e8bae31b0489b42328fb07f7d8b6f42647"} Oct 01 10:38:40 crc kubenswrapper[4787]: I1001 10:38:40.416464 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x2bwg/must-gather-ht4gk" event={"ID":"32eecbc7-b51a-465f-84ce-5352da7cf52a","Type":"ContainerStarted","Data":"d5bdcdb73c47397f098127324e8996a20cb0e14f524c114cdb1b4dc0ebc11b08"} Oct 01 10:38:40 crc kubenswrapper[4787]: I1001 10:38:40.443827 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-x2bwg/must-gather-ht4gk" podStartSLOduration=4.443808605 podStartE2EDuration="4.443808605s" podCreationTimestamp="2025-10-01 10:38:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 10:38:40.434125867 +0000 UTC m=+3752.549270034" watchObservedRunningTime="2025-10-01 10:38:40.443808605 +0000 UTC m=+3752.558952762" Oct 01 10:38:43 crc kubenswrapper[4787]: I1001 10:38:43.523734 4787 scope.go:117] "RemoveContainer" containerID="7cb2890ef30bbede4935afca82be38385def845a6f5b1c0bd5934ac785342b56" Oct 01 10:38:43 crc kubenswrapper[4787]: E1001 10:38:43.524483 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:38:44 crc kubenswrapper[4787]: I1001 10:38:44.050327 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-x2bwg/crc-debug-mc5mx"] Oct 01 10:38:44 crc kubenswrapper[4787]: I1001 10:38:44.051725 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x2bwg/crc-debug-mc5mx" Oct 01 10:38:44 crc kubenswrapper[4787]: I1001 10:38:44.055313 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-x2bwg"/"default-dockercfg-dnxsc" Oct 01 10:38:44 crc kubenswrapper[4787]: I1001 10:38:44.189070 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mb7jv\" (UniqueName: \"kubernetes.io/projected/a82e51ac-c976-4412-85ab-875ed39ba40d-kube-api-access-mb7jv\") pod \"crc-debug-mc5mx\" (UID: \"a82e51ac-c976-4412-85ab-875ed39ba40d\") " pod="openshift-must-gather-x2bwg/crc-debug-mc5mx" Oct 01 10:38:44 crc kubenswrapper[4787]: I1001 10:38:44.189261 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a82e51ac-c976-4412-85ab-875ed39ba40d-host\") pod \"crc-debug-mc5mx\" (UID: \"a82e51ac-c976-4412-85ab-875ed39ba40d\") " pod="openshift-must-gather-x2bwg/crc-debug-mc5mx" Oct 01 10:38:44 crc kubenswrapper[4787]: I1001 10:38:44.290773 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mb7jv\" (UniqueName: \"kubernetes.io/projected/a82e51ac-c976-4412-85ab-875ed39ba40d-kube-api-access-mb7jv\") pod \"crc-debug-mc5mx\" (UID: \"a82e51ac-c976-4412-85ab-875ed39ba40d\") " pod="openshift-must-gather-x2bwg/crc-debug-mc5mx" Oct 01 10:38:44 crc kubenswrapper[4787]: I1001 10:38:44.290920 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a82e51ac-c976-4412-85ab-875ed39ba40d-host\") pod \"crc-debug-mc5mx\" (UID: \"a82e51ac-c976-4412-85ab-875ed39ba40d\") " pod="openshift-must-gather-x2bwg/crc-debug-mc5mx" Oct 01 10:38:44 crc kubenswrapper[4787]: I1001 10:38:44.291280 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a82e51ac-c976-4412-85ab-875ed39ba40d-host\") pod \"crc-debug-mc5mx\" (UID: \"a82e51ac-c976-4412-85ab-875ed39ba40d\") " pod="openshift-must-gather-x2bwg/crc-debug-mc5mx" Oct 01 10:38:44 crc kubenswrapper[4787]: I1001 10:38:44.338835 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mb7jv\" (UniqueName: \"kubernetes.io/projected/a82e51ac-c976-4412-85ab-875ed39ba40d-kube-api-access-mb7jv\") pod \"crc-debug-mc5mx\" (UID: \"a82e51ac-c976-4412-85ab-875ed39ba40d\") " pod="openshift-must-gather-x2bwg/crc-debug-mc5mx" Oct 01 10:38:44 crc kubenswrapper[4787]: I1001 10:38:44.382357 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x2bwg/crc-debug-mc5mx" Oct 01 10:38:44 crc kubenswrapper[4787]: I1001 10:38:44.481038 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x2bwg/crc-debug-mc5mx" event={"ID":"a82e51ac-c976-4412-85ab-875ed39ba40d","Type":"ContainerStarted","Data":"2090b60483a843b5187c1d50a89c76a6268b8a6da5375371f5b6ce6b8ad3b420"} Oct 01 10:38:45 crc kubenswrapper[4787]: I1001 10:38:45.491280 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x2bwg/crc-debug-mc5mx" event={"ID":"a82e51ac-c976-4412-85ab-875ed39ba40d","Type":"ContainerStarted","Data":"3c0a8d063a825052343d120852a4586d34169cfae7fa50df892b915a15b118ec"} Oct 01 10:38:45 crc kubenswrapper[4787]: I1001 10:38:45.508521 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-x2bwg/crc-debug-mc5mx" podStartSLOduration=1.5084970709999999 podStartE2EDuration="1.508497071s" podCreationTimestamp="2025-10-01 10:38:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 10:38:45.504442482 +0000 UTC m=+3757.619586649" watchObservedRunningTime="2025-10-01 10:38:45.508497071 +0000 UTC m=+3757.623641238" Oct 01 10:38:56 crc kubenswrapper[4787]: I1001 10:38:56.523906 4787 scope.go:117] "RemoveContainer" containerID="7cb2890ef30bbede4935afca82be38385def845a6f5b1c0bd5934ac785342b56" Oct 01 10:38:56 crc kubenswrapper[4787]: E1001 10:38:56.524599 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:39:11 crc kubenswrapper[4787]: I1001 10:39:11.524068 4787 scope.go:117] "RemoveContainer" containerID="7cb2890ef30bbede4935afca82be38385def845a6f5b1c0bd5934ac785342b56" Oct 01 10:39:11 crc kubenswrapper[4787]: E1001 10:39:11.525088 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:39:23 crc kubenswrapper[4787]: I1001 10:39:23.526195 4787 scope.go:117] "RemoveContainer" containerID="7cb2890ef30bbede4935afca82be38385def845a6f5b1c0bd5934ac785342b56" Oct 01 10:39:23 crc kubenswrapper[4787]: E1001 10:39:23.527637 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:39:34 crc kubenswrapper[4787]: I1001 10:39:34.529657 4787 scope.go:117] "RemoveContainer" containerID="7cb2890ef30bbede4935afca82be38385def845a6f5b1c0bd5934ac785342b56" Oct 01 10:39:34 crc kubenswrapper[4787]: E1001 10:39:34.530412 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:39:47 crc kubenswrapper[4787]: I1001 10:39:47.524678 4787 scope.go:117] "RemoveContainer" containerID="7cb2890ef30bbede4935afca82be38385def845a6f5b1c0bd5934ac785342b56" Oct 01 10:39:47 crc kubenswrapper[4787]: E1001 10:39:47.525525 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:39:48 crc kubenswrapper[4787]: I1001 10:39:48.996893 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-887c78f44-f74pm_873541c7-e7e7-4219-8ce0-c8d5edfbe308/barbican-api/0.log" Oct 01 10:39:49 crc kubenswrapper[4787]: I1001 10:39:49.009306 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-887c78f44-f74pm_873541c7-e7e7-4219-8ce0-c8d5edfbe308/barbican-api-log/0.log" Oct 01 10:39:49 crc kubenswrapper[4787]: I1001 10:39:49.213903 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-699b4d5c4-jdg79_484a5c7d-7633-4e34-a235-af316bbccada/barbican-keystone-listener-log/0.log" Oct 01 10:39:49 crc kubenswrapper[4787]: I1001 10:39:49.244892 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-699b4d5c4-jdg79_484a5c7d-7633-4e34-a235-af316bbccada/barbican-keystone-listener/0.log" Oct 01 10:39:49 crc kubenswrapper[4787]: I1001 10:39:49.468469 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-64d9c8b8f9-zv7w7_ea69f2cb-9d4f-4b3c-a0ba-8a3fc8fce274/barbican-worker/0.log" Oct 01 10:39:49 crc kubenswrapper[4787]: I1001 10:39:49.502054 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-64d9c8b8f9-zv7w7_ea69f2cb-9d4f-4b3c-a0ba-8a3fc8fce274/barbican-worker-log/0.log" Oct 01 10:39:49 crc kubenswrapper[4787]: I1001 10:39:49.726320 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-s4tf6_d1018ece-ee61-4913-bd2f-64f996ef135c/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 10:39:49 crc kubenswrapper[4787]: I1001 10:39:49.992014 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0cdb08fb-0a31-40dd-98c6-d13926671f45/ceilometer-central-agent/0.log" Oct 01 10:39:50 crc kubenswrapper[4787]: I1001 10:39:50.002048 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0cdb08fb-0a31-40dd-98c6-d13926671f45/ceilometer-notification-agent/0.log" Oct 01 10:39:50 crc kubenswrapper[4787]: I1001 10:39:50.057323 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0cdb08fb-0a31-40dd-98c6-d13926671f45/proxy-httpd/0.log" Oct 01 10:39:50 crc kubenswrapper[4787]: I1001 10:39:50.201794 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0cdb08fb-0a31-40dd-98c6-d13926671f45/sg-core/0.log" Oct 01 10:39:50 crc kubenswrapper[4787]: I1001 10:39:50.320556 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_a05bd18f-4272-4ff1-ac85-65f0a20cbdcf/cinder-api/0.log" Oct 01 10:39:50 crc kubenswrapper[4787]: I1001 10:39:50.454102 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_a05bd18f-4272-4ff1-ac85-65f0a20cbdcf/cinder-api-log/0.log" Oct 01 10:39:50 crc kubenswrapper[4787]: I1001 10:39:50.599991 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_8a844181-c2d3-4755-a0e2-d3706a9eb3e0/cinder-scheduler/0.log" Oct 01 10:39:50 crc kubenswrapper[4787]: I1001 10:39:50.696591 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_8a844181-c2d3-4755-a0e2-d3706a9eb3e0/probe/0.log" Oct 01 10:39:50 crc kubenswrapper[4787]: I1001 10:39:50.829696 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-llpfd_8dd2f969-9904-4848-b0e5-f852a8b958ea/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 10:39:51 crc kubenswrapper[4787]: I1001 10:39:51.050576 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-ddl72_def7d40c-8191-4256-b8ad-4fe9272d5fae/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 10:39:51 crc kubenswrapper[4787]: I1001 10:39:51.290400 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-zwc7x_c2b165b4-e2e9-4b0c-91f7-db75bcbc113d/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 10:39:51 crc kubenswrapper[4787]: I1001 10:39:51.331161 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6966dc867f-cvjjh_fc33ac30-efd1-438d-9ab5-d700dfd27efe/init/0.log" Oct 01 10:39:51 crc kubenswrapper[4787]: I1001 10:39:51.498357 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6966dc867f-cvjjh_fc33ac30-efd1-438d-9ab5-d700dfd27efe/init/0.log" Oct 01 10:39:51 crc kubenswrapper[4787]: I1001 10:39:51.583203 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6966dc867f-cvjjh_fc33ac30-efd1-438d-9ab5-d700dfd27efe/dnsmasq-dns/0.log" Oct 01 10:39:51 crc kubenswrapper[4787]: I1001 10:39:51.736669 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-4jp68_f70ffa29-bf25-44a2-bd45-822e5f59a5d5/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 10:39:51 crc kubenswrapper[4787]: I1001 10:39:51.795595 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_137125b0-e356-4dbc-be53-2dbd3a06825f/glance-httpd/0.log" Oct 01 10:39:51 crc kubenswrapper[4787]: I1001 10:39:51.977355 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_137125b0-e356-4dbc-be53-2dbd3a06825f/glance-log/0.log" Oct 01 10:39:52 crc kubenswrapper[4787]: I1001 10:39:52.120590 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_ddca8826-066e-45d5-9536-59ef0bfc8c61/glance-httpd/0.log" Oct 01 10:39:52 crc kubenswrapper[4787]: I1001 10:39:52.180554 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_ddca8826-066e-45d5-9536-59ef0bfc8c61/glance-log/0.log" Oct 01 10:39:52 crc kubenswrapper[4787]: I1001 10:39:52.487763 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7d65d5957b-44rtv_db1d14ba-8d00-4508-9c89-fdadb562ad0d/horizon/0.log" Oct 01 10:39:52 crc kubenswrapper[4787]: I1001 10:39:52.509790 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-79skv_d7c84e3c-e5bc-46e7-8288-f68b75bd8ab1/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 10:39:52 crc kubenswrapper[4787]: I1001 10:39:52.777125 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-qvjjs_654cfd15-20ca-4754-9de7-ebe67905c8d2/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 10:39:52 crc kubenswrapper[4787]: I1001 10:39:52.869523 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7d65d5957b-44rtv_db1d14ba-8d00-4508-9c89-fdadb562ad0d/horizon-log/0.log" Oct 01 10:39:53 crc kubenswrapper[4787]: I1001 10:39:53.157750 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29321881-dnxcs_7b0f7d9e-ab06-41c1-bac1-54004f737044/keystone-cron/0.log" Oct 01 10:39:53 crc kubenswrapper[4787]: I1001 10:39:53.245086 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-7f57487985-p72bt_db5966fa-e6ce-4fd2-9a2c-b17671db2625/keystone-api/0.log" Oct 01 10:39:53 crc kubenswrapper[4787]: I1001 10:39:53.275438 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_c1c9f3e0-fa6b-4ef4-ae4a-a0aea3bfc262/kube-state-metrics/0.log" Oct 01 10:39:53 crc kubenswrapper[4787]: I1001 10:39:53.396313 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-d4wlr_c8a427c3-8fbf-487f-abf5-ad5d2ec0d89c/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 10:39:53 crc kubenswrapper[4787]: I1001 10:39:53.819019 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-58d9cd74f7-86sts_1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0/neutron-api/0.log" Oct 01 10:39:53 crc kubenswrapper[4787]: I1001 10:39:53.860251 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-58d9cd74f7-86sts_1e4d70a9-10b1-4bb7-a151-56f16bb2cbe0/neutron-httpd/0.log" Oct 01 10:39:54 crc kubenswrapper[4787]: I1001 10:39:54.039573 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-jkdl6_e7fb97de-2759-41c1-b090-99ed2c95e92c/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 10:39:54 crc kubenswrapper[4787]: I1001 10:39:54.598879 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_59257c48-7bb8-4aad-9f6c-71d4459f2fb7/nova-api-log/0.log" Oct 01 10:39:54 crc kubenswrapper[4787]: I1001 10:39:54.945317 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_39f3448f-eee4-419b-a33b-8776fc5df599/nova-cell0-conductor-conductor/0.log" Oct 01 10:39:55 crc kubenswrapper[4787]: I1001 10:39:55.090640 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_59257c48-7bb8-4aad-9f6c-71d4459f2fb7/nova-api-api/0.log" Oct 01 10:39:55 crc kubenswrapper[4787]: I1001 10:39:55.272311 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_1f4a94ee-d626-48f8-93c2-547212f23da4/nova-cell1-conductor-conductor/0.log" Oct 01 10:39:55 crc kubenswrapper[4787]: I1001 10:39:55.487672 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_f8b719ce-ceea-4c80-8c8d-846f2bf7a402/nova-cell1-novncproxy-novncproxy/0.log" Oct 01 10:39:55 crc kubenswrapper[4787]: I1001 10:39:55.565800 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-ltfkq_4f22f28b-714f-4ab7-a90b-a400a500e810/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 10:39:55 crc kubenswrapper[4787]: I1001 10:39:55.886035 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f3dac849-090f-4eaa-bd8a-11a9e85f0317/nova-metadata-log/0.log" Oct 01 10:39:56 crc kubenswrapper[4787]: I1001 10:39:56.360007 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_35a5073c-d1ab-4fa9-b1c5-3a6df1d42bcc/nova-scheduler-scheduler/0.log" Oct 01 10:39:56 crc kubenswrapper[4787]: I1001 10:39:56.611308 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7924a2ac-f411-42d6-8b23-f22faf58351a/mysql-bootstrap/0.log" Oct 01 10:39:56 crc kubenswrapper[4787]: I1001 10:39:56.747974 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7924a2ac-f411-42d6-8b23-f22faf58351a/mysql-bootstrap/0.log" Oct 01 10:39:56 crc kubenswrapper[4787]: I1001 10:39:56.873788 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7924a2ac-f411-42d6-8b23-f22faf58351a/galera/0.log" Oct 01 10:39:57 crc kubenswrapper[4787]: I1001 10:39:57.088041 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c/mysql-bootstrap/0.log" Oct 01 10:39:57 crc kubenswrapper[4787]: I1001 10:39:57.378922 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c/mysql-bootstrap/0.log" Oct 01 10:39:57 crc kubenswrapper[4787]: I1001 10:39:57.394424 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_37a36e0e-c7f4-4ed8-a21d-e0b09741aa4c/galera/0.log" Oct 01 10:39:57 crc kubenswrapper[4787]: I1001 10:39:57.496888 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f3dac849-090f-4eaa-bd8a-11a9e85f0317/nova-metadata-metadata/0.log" Oct 01 10:39:57 crc kubenswrapper[4787]: I1001 10:39:57.632585 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_ccb620ed-ad28-4ec7-8319-6d5bb3ce10ec/openstackclient/0.log" Oct 01 10:39:57 crc kubenswrapper[4787]: I1001 10:39:57.903532 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-l8fqd_36eb99a8-609a-4af2-9fa7-d0051806659e/ovn-controller/0.log" Oct 01 10:39:58 crc kubenswrapper[4787]: I1001 10:39:58.002674 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-d9cxp_11dbf4b0-9694-40e2-a122-9da928c50ad8/openstack-network-exporter/0.log" Oct 01 10:39:58 crc kubenswrapper[4787]: I1001 10:39:58.181455 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6htcd_948a10c7-fb61-4110-a08e-b2fd42cb842a/ovsdb-server-init/0.log" Oct 01 10:39:58 crc kubenswrapper[4787]: I1001 10:39:58.394198 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6htcd_948a10c7-fb61-4110-a08e-b2fd42cb842a/ovsdb-server-init/0.log" Oct 01 10:39:58 crc kubenswrapper[4787]: I1001 10:39:58.411328 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6htcd_948a10c7-fb61-4110-a08e-b2fd42cb842a/ovs-vswitchd/0.log" Oct 01 10:39:58 crc kubenswrapper[4787]: I1001 10:39:58.426387 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6htcd_948a10c7-fb61-4110-a08e-b2fd42cb842a/ovsdb-server/0.log" Oct 01 10:39:58 crc kubenswrapper[4787]: I1001 10:39:58.687969 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-htlbc_dc18c70f-f7c7-4a60-87e6-699320d382fc/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 10:39:58 crc kubenswrapper[4787]: I1001 10:39:58.875217 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_e5f4597a-1f01-4082-aada-cc7b976fad9a/openstack-network-exporter/0.log" Oct 01 10:39:58 crc kubenswrapper[4787]: I1001 10:39:58.913904 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_e5f4597a-1f01-4082-aada-cc7b976fad9a/ovn-northd/0.log" Oct 01 10:39:59 crc kubenswrapper[4787]: I1001 10:39:59.126176 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_ee50aeec-8875-45c2-80e2-7121480ae89f/openstack-network-exporter/0.log" Oct 01 10:39:59 crc kubenswrapper[4787]: I1001 10:39:59.146304 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_ee50aeec-8875-45c2-80e2-7121480ae89f/ovsdbserver-nb/0.log" Oct 01 10:39:59 crc kubenswrapper[4787]: I1001 10:39:59.344820 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_696f6d7f-812c-4d36-a799-397b83edf291/openstack-network-exporter/0.log" Oct 01 10:39:59 crc kubenswrapper[4787]: I1001 10:39:59.403232 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_696f6d7f-812c-4d36-a799-397b83edf291/ovsdbserver-sb/0.log" Oct 01 10:39:59 crc kubenswrapper[4787]: I1001 10:39:59.641700 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-84d5f4684d-p2xjn_a617751d-ce49-4357-bed0-32a3f63d4d84/placement-api/0.log" Oct 01 10:39:59 crc kubenswrapper[4787]: I1001 10:39:59.772546 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-84d5f4684d-p2xjn_a617751d-ce49-4357-bed0-32a3f63d4d84/placement-log/0.log" Oct 01 10:39:59 crc kubenswrapper[4787]: I1001 10:39:59.927508 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_d0353483-c978-4d9a-9c34-73ee379325ba/setup-container/0.log" Oct 01 10:40:00 crc kubenswrapper[4787]: I1001 10:40:00.086388 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_d0353483-c978-4d9a-9c34-73ee379325ba/setup-container/0.log" Oct 01 10:40:00 crc kubenswrapper[4787]: I1001 10:40:00.231364 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_d0353483-c978-4d9a-9c34-73ee379325ba/rabbitmq/0.log" Oct 01 10:40:00 crc kubenswrapper[4787]: I1001 10:40:00.343269 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_eb5022f8-276b-4d6d-917b-be96a1d993bf/setup-container/0.log" Oct 01 10:40:00 crc kubenswrapper[4787]: I1001 10:40:00.517688 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_eb5022f8-276b-4d6d-917b-be96a1d993bf/setup-container/0.log" Oct 01 10:40:00 crc kubenswrapper[4787]: I1001 10:40:00.544319 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_eb5022f8-276b-4d6d-917b-be96a1d993bf/rabbitmq/0.log" Oct 01 10:40:00 crc kubenswrapper[4787]: I1001 10:40:00.737090 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-llnrf_728521d2-1e71-4da3-a8bc-bd68e02eae35/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 10:40:00 crc kubenswrapper[4787]: I1001 10:40:00.891418 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-j5ltq_8d46852a-ab64-4e0c-b29a-e9bde8305a94/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 10:40:01 crc kubenswrapper[4787]: I1001 10:40:01.064067 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-wk6cx_d8ba2dbe-8a68-4da2-ac35-664a50fb5061/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 10:40:01 crc kubenswrapper[4787]: I1001 10:40:01.316009 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-ws2cx_389d6832-2d3e-49a0-afbd-f88359db6324/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 10:40:01 crc kubenswrapper[4787]: I1001 10:40:01.430511 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-rhhxm_49e3ae1d-7abd-4e22-a333-54e10db349a6/ssh-known-hosts-edpm-deployment/0.log" Oct 01 10:40:01 crc kubenswrapper[4787]: I1001 10:40:01.523490 4787 scope.go:117] "RemoveContainer" containerID="7cb2890ef30bbede4935afca82be38385def845a6f5b1c0bd5934ac785342b56" Oct 01 10:40:01 crc kubenswrapper[4787]: E1001 10:40:01.523746 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8vh6d_openshift-machine-config-operator(fa51b883-5f3e-4141-9d4c-37704aac7718)\"" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" Oct 01 10:40:01 crc kubenswrapper[4787]: I1001 10:40:01.706446 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5c7b59dd4f-spxm4_e743cd5e-e8c6-4fe2-9480-3a30316b8e23/proxy-server/0.log" Oct 01 10:40:01 crc kubenswrapper[4787]: I1001 10:40:01.741522 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5c7b59dd4f-spxm4_e743cd5e-e8c6-4fe2-9480-3a30316b8e23/proxy-httpd/0.log" Oct 01 10:40:01 crc kubenswrapper[4787]: I1001 10:40:01.895670 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-kkdkg_ed714165-98db-45a6-b96b-69601bff3e56/swift-ring-rebalance/0.log" Oct 01 10:40:02 crc kubenswrapper[4787]: I1001 10:40:02.155322 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3432bc16-99b4-431a-aeb1-600e826bbc3e/account-auditor/0.log" Oct 01 10:40:02 crc kubenswrapper[4787]: I1001 10:40:02.160889 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3432bc16-99b4-431a-aeb1-600e826bbc3e/account-reaper/0.log" Oct 01 10:40:02 crc kubenswrapper[4787]: I1001 10:40:02.340369 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3432bc16-99b4-431a-aeb1-600e826bbc3e/account-replicator/0.log" Oct 01 10:40:02 crc kubenswrapper[4787]: I1001 10:40:02.379438 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3432bc16-99b4-431a-aeb1-600e826bbc3e/container-auditor/0.log" Oct 01 10:40:02 crc kubenswrapper[4787]: I1001 10:40:02.415934 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3432bc16-99b4-431a-aeb1-600e826bbc3e/account-server/0.log" Oct 01 10:40:02 crc kubenswrapper[4787]: I1001 10:40:02.577649 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3432bc16-99b4-431a-aeb1-600e826bbc3e/container-server/0.log" Oct 01 10:40:02 crc kubenswrapper[4787]: I1001 10:40:02.657225 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3432bc16-99b4-431a-aeb1-600e826bbc3e/container-updater/0.log" Oct 01 10:40:02 crc kubenswrapper[4787]: I1001 10:40:02.681617 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3432bc16-99b4-431a-aeb1-600e826bbc3e/container-replicator/0.log" Oct 01 10:40:02 crc kubenswrapper[4787]: I1001 10:40:02.828709 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3432bc16-99b4-431a-aeb1-600e826bbc3e/object-auditor/0.log" Oct 01 10:40:02 crc kubenswrapper[4787]: I1001 10:40:02.873891 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3432bc16-99b4-431a-aeb1-600e826bbc3e/object-expirer/0.log" Oct 01 10:40:02 crc kubenswrapper[4787]: I1001 10:40:02.988056 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3432bc16-99b4-431a-aeb1-600e826bbc3e/object-replicator/0.log" Oct 01 10:40:03 crc kubenswrapper[4787]: I1001 10:40:03.057128 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3432bc16-99b4-431a-aeb1-600e826bbc3e/object-server/0.log" Oct 01 10:40:03 crc kubenswrapper[4787]: I1001 10:40:03.195706 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3432bc16-99b4-431a-aeb1-600e826bbc3e/object-updater/0.log" Oct 01 10:40:03 crc kubenswrapper[4787]: I1001 10:40:03.221261 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3432bc16-99b4-431a-aeb1-600e826bbc3e/rsync/0.log" Oct 01 10:40:03 crc kubenswrapper[4787]: I1001 10:40:03.310511 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3432bc16-99b4-431a-aeb1-600e826bbc3e/swift-recon-cron/0.log" Oct 01 10:40:03 crc kubenswrapper[4787]: I1001 10:40:03.471443 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-gxbs9_56b1c51b-bdcd-4c55-a00d-105ab4bc04b0/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 10:40:03 crc kubenswrapper[4787]: I1001 10:40:03.698630 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_4b136104-8797-40d5-9ea1-fec8c6be0e90/tempest-tests-tempest-tests-runner/0.log" Oct 01 10:40:03 crc kubenswrapper[4787]: I1001 10:40:03.833911 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_9aaa5922-539c-4e98-99ac-a08a31b14d53/test-operator-logs-container/0.log" Oct 01 10:40:03 crc kubenswrapper[4787]: I1001 10:40:03.967486 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-nkbqd_08369955-6564-47c7-bed0-64893e898e1b/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 01 10:40:12 crc kubenswrapper[4787]: I1001 10:40:12.984385 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_7137d092-ff92-4b8d-84c5-9d9d269ddcc7/memcached/0.log" Oct 01 10:40:14 crc kubenswrapper[4787]: I1001 10:40:14.524998 4787 scope.go:117] "RemoveContainer" containerID="7cb2890ef30bbede4935afca82be38385def845a6f5b1c0bd5934ac785342b56" Oct 01 10:40:15 crc kubenswrapper[4787]: I1001 10:40:15.425500 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" event={"ID":"fa51b883-5f3e-4141-9d4c-37704aac7718","Type":"ContainerStarted","Data":"6e2358915272c02486736120944ef044ca65ab4dad3e49439ebe14ee8e2531f0"} Oct 01 10:40:33 crc kubenswrapper[4787]: I1001 10:40:33.631031 4787 scope.go:117] "RemoveContainer" containerID="dd019f5093eb2e44f07d38ecec1f9ff118048c1e7532ea4677715101811b8bcd" Oct 01 10:40:39 crc kubenswrapper[4787]: I1001 10:40:39.649568 4787 generic.go:334] "Generic (PLEG): container finished" podID="a82e51ac-c976-4412-85ab-875ed39ba40d" containerID="3c0a8d063a825052343d120852a4586d34169cfae7fa50df892b915a15b118ec" exitCode=0 Oct 01 10:40:39 crc kubenswrapper[4787]: I1001 10:40:39.649769 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x2bwg/crc-debug-mc5mx" event={"ID":"a82e51ac-c976-4412-85ab-875ed39ba40d","Type":"ContainerDied","Data":"3c0a8d063a825052343d120852a4586d34169cfae7fa50df892b915a15b118ec"} Oct 01 10:40:40 crc kubenswrapper[4787]: I1001 10:40:40.766230 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x2bwg/crc-debug-mc5mx" Oct 01 10:40:40 crc kubenswrapper[4787]: I1001 10:40:40.805892 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-x2bwg/crc-debug-mc5mx"] Oct 01 10:40:40 crc kubenswrapper[4787]: I1001 10:40:40.814862 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-x2bwg/crc-debug-mc5mx"] Oct 01 10:40:40 crc kubenswrapper[4787]: I1001 10:40:40.930220 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mb7jv\" (UniqueName: \"kubernetes.io/projected/a82e51ac-c976-4412-85ab-875ed39ba40d-kube-api-access-mb7jv\") pod \"a82e51ac-c976-4412-85ab-875ed39ba40d\" (UID: \"a82e51ac-c976-4412-85ab-875ed39ba40d\") " Oct 01 10:40:40 crc kubenswrapper[4787]: I1001 10:40:40.930496 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a82e51ac-c976-4412-85ab-875ed39ba40d-host\") pod \"a82e51ac-c976-4412-85ab-875ed39ba40d\" (UID: \"a82e51ac-c976-4412-85ab-875ed39ba40d\") " Oct 01 10:40:40 crc kubenswrapper[4787]: I1001 10:40:40.930625 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a82e51ac-c976-4412-85ab-875ed39ba40d-host" (OuterVolumeSpecName: "host") pod "a82e51ac-c976-4412-85ab-875ed39ba40d" (UID: "a82e51ac-c976-4412-85ab-875ed39ba40d"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 10:40:40 crc kubenswrapper[4787]: I1001 10:40:40.931533 4787 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a82e51ac-c976-4412-85ab-875ed39ba40d-host\") on node \"crc\" DevicePath \"\"" Oct 01 10:40:40 crc kubenswrapper[4787]: I1001 10:40:40.936324 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a82e51ac-c976-4412-85ab-875ed39ba40d-kube-api-access-mb7jv" (OuterVolumeSpecName: "kube-api-access-mb7jv") pod "a82e51ac-c976-4412-85ab-875ed39ba40d" (UID: "a82e51ac-c976-4412-85ab-875ed39ba40d"). InnerVolumeSpecName "kube-api-access-mb7jv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:40:41 crc kubenswrapper[4787]: I1001 10:40:41.033155 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mb7jv\" (UniqueName: \"kubernetes.io/projected/a82e51ac-c976-4412-85ab-875ed39ba40d-kube-api-access-mb7jv\") on node \"crc\" DevicePath \"\"" Oct 01 10:40:41 crc kubenswrapper[4787]: I1001 10:40:41.678940 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x2bwg/crc-debug-mc5mx" Oct 01 10:40:41 crc kubenswrapper[4787]: I1001 10:40:41.679554 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2090b60483a843b5187c1d50a89c76a6268b8a6da5375371f5b6ce6b8ad3b420" Oct 01 10:40:41 crc kubenswrapper[4787]: I1001 10:40:41.975784 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-x2bwg/crc-debug-2sbvd"] Oct 01 10:40:41 crc kubenswrapper[4787]: E1001 10:40:41.976238 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a82e51ac-c976-4412-85ab-875ed39ba40d" containerName="container-00" Oct 01 10:40:41 crc kubenswrapper[4787]: I1001 10:40:41.976251 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="a82e51ac-c976-4412-85ab-875ed39ba40d" containerName="container-00" Oct 01 10:40:41 crc kubenswrapper[4787]: I1001 10:40:41.976470 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="a82e51ac-c976-4412-85ab-875ed39ba40d" containerName="container-00" Oct 01 10:40:41 crc kubenswrapper[4787]: I1001 10:40:41.977105 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x2bwg/crc-debug-2sbvd" Oct 01 10:40:41 crc kubenswrapper[4787]: I1001 10:40:41.978655 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-x2bwg"/"default-dockercfg-dnxsc" Oct 01 10:40:42 crc kubenswrapper[4787]: I1001 10:40:42.051191 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6crg9\" (UniqueName: \"kubernetes.io/projected/f4388fd5-b907-4c60-aa8f-e39f6a5ae2e4-kube-api-access-6crg9\") pod \"crc-debug-2sbvd\" (UID: \"f4388fd5-b907-4c60-aa8f-e39f6a5ae2e4\") " pod="openshift-must-gather-x2bwg/crc-debug-2sbvd" Oct 01 10:40:42 crc kubenswrapper[4787]: I1001 10:40:42.051496 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f4388fd5-b907-4c60-aa8f-e39f6a5ae2e4-host\") pod \"crc-debug-2sbvd\" (UID: \"f4388fd5-b907-4c60-aa8f-e39f6a5ae2e4\") " pod="openshift-must-gather-x2bwg/crc-debug-2sbvd" Oct 01 10:40:42 crc kubenswrapper[4787]: I1001 10:40:42.153846 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f4388fd5-b907-4c60-aa8f-e39f6a5ae2e4-host\") pod \"crc-debug-2sbvd\" (UID: \"f4388fd5-b907-4c60-aa8f-e39f6a5ae2e4\") " pod="openshift-must-gather-x2bwg/crc-debug-2sbvd" Oct 01 10:40:42 crc kubenswrapper[4787]: I1001 10:40:42.153973 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f4388fd5-b907-4c60-aa8f-e39f6a5ae2e4-host\") pod \"crc-debug-2sbvd\" (UID: \"f4388fd5-b907-4c60-aa8f-e39f6a5ae2e4\") " pod="openshift-must-gather-x2bwg/crc-debug-2sbvd" Oct 01 10:40:42 crc kubenswrapper[4787]: I1001 10:40:42.154176 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6crg9\" (UniqueName: \"kubernetes.io/projected/f4388fd5-b907-4c60-aa8f-e39f6a5ae2e4-kube-api-access-6crg9\") pod \"crc-debug-2sbvd\" (UID: \"f4388fd5-b907-4c60-aa8f-e39f6a5ae2e4\") " pod="openshift-must-gather-x2bwg/crc-debug-2sbvd" Oct 01 10:40:42 crc kubenswrapper[4787]: I1001 10:40:42.172209 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6crg9\" (UniqueName: \"kubernetes.io/projected/f4388fd5-b907-4c60-aa8f-e39f6a5ae2e4-kube-api-access-6crg9\") pod \"crc-debug-2sbvd\" (UID: \"f4388fd5-b907-4c60-aa8f-e39f6a5ae2e4\") " pod="openshift-must-gather-x2bwg/crc-debug-2sbvd" Oct 01 10:40:42 crc kubenswrapper[4787]: I1001 10:40:42.296039 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x2bwg/crc-debug-2sbvd" Oct 01 10:40:42 crc kubenswrapper[4787]: I1001 10:40:42.537860 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a82e51ac-c976-4412-85ab-875ed39ba40d" path="/var/lib/kubelet/pods/a82e51ac-c976-4412-85ab-875ed39ba40d/volumes" Oct 01 10:40:42 crc kubenswrapper[4787]: I1001 10:40:42.680670 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x2bwg/crc-debug-2sbvd" event={"ID":"f4388fd5-b907-4c60-aa8f-e39f6a5ae2e4","Type":"ContainerStarted","Data":"75899f576cc094a662b947821d937bfed0459104202add7eb25ebf2fd66290c3"} Oct 01 10:40:42 crc kubenswrapper[4787]: I1001 10:40:42.680719 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x2bwg/crc-debug-2sbvd" event={"ID":"f4388fd5-b907-4c60-aa8f-e39f6a5ae2e4","Type":"ContainerStarted","Data":"6c46935dfc876fbc6464b7ab220ff36dcd90e19dda873549257c553cc029fff0"} Oct 01 10:40:42 crc kubenswrapper[4787]: I1001 10:40:42.698372 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-x2bwg/crc-debug-2sbvd" podStartSLOduration=1.6983470330000001 podStartE2EDuration="1.698347033s" podCreationTimestamp="2025-10-01 10:40:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 10:40:42.691969857 +0000 UTC m=+3874.807114014" watchObservedRunningTime="2025-10-01 10:40:42.698347033 +0000 UTC m=+3874.813491200" Oct 01 10:40:43 crc kubenswrapper[4787]: I1001 10:40:43.690260 4787 generic.go:334] "Generic (PLEG): container finished" podID="f4388fd5-b907-4c60-aa8f-e39f6a5ae2e4" containerID="75899f576cc094a662b947821d937bfed0459104202add7eb25ebf2fd66290c3" exitCode=0 Oct 01 10:40:43 crc kubenswrapper[4787]: I1001 10:40:43.690385 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x2bwg/crc-debug-2sbvd" event={"ID":"f4388fd5-b907-4c60-aa8f-e39f6a5ae2e4","Type":"ContainerDied","Data":"75899f576cc094a662b947821d937bfed0459104202add7eb25ebf2fd66290c3"} Oct 01 10:40:44 crc kubenswrapper[4787]: I1001 10:40:44.827905 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x2bwg/crc-debug-2sbvd" Oct 01 10:40:44 crc kubenswrapper[4787]: I1001 10:40:44.900157 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6crg9\" (UniqueName: \"kubernetes.io/projected/f4388fd5-b907-4c60-aa8f-e39f6a5ae2e4-kube-api-access-6crg9\") pod \"f4388fd5-b907-4c60-aa8f-e39f6a5ae2e4\" (UID: \"f4388fd5-b907-4c60-aa8f-e39f6a5ae2e4\") " Oct 01 10:40:44 crc kubenswrapper[4787]: I1001 10:40:44.900370 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f4388fd5-b907-4c60-aa8f-e39f6a5ae2e4-host\") pod \"f4388fd5-b907-4c60-aa8f-e39f6a5ae2e4\" (UID: \"f4388fd5-b907-4c60-aa8f-e39f6a5ae2e4\") " Oct 01 10:40:44 crc kubenswrapper[4787]: I1001 10:40:44.900534 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4388fd5-b907-4c60-aa8f-e39f6a5ae2e4-host" (OuterVolumeSpecName: "host") pod "f4388fd5-b907-4c60-aa8f-e39f6a5ae2e4" (UID: "f4388fd5-b907-4c60-aa8f-e39f6a5ae2e4"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 10:40:44 crc kubenswrapper[4787]: I1001 10:40:44.901670 4787 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f4388fd5-b907-4c60-aa8f-e39f6a5ae2e4-host\") on node \"crc\" DevicePath \"\"" Oct 01 10:40:44 crc kubenswrapper[4787]: I1001 10:40:44.907670 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4388fd5-b907-4c60-aa8f-e39f6a5ae2e4-kube-api-access-6crg9" (OuterVolumeSpecName: "kube-api-access-6crg9") pod "f4388fd5-b907-4c60-aa8f-e39f6a5ae2e4" (UID: "f4388fd5-b907-4c60-aa8f-e39f6a5ae2e4"). InnerVolumeSpecName "kube-api-access-6crg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:40:45 crc kubenswrapper[4787]: I1001 10:40:45.003102 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6crg9\" (UniqueName: \"kubernetes.io/projected/f4388fd5-b907-4c60-aa8f-e39f6a5ae2e4-kube-api-access-6crg9\") on node \"crc\" DevicePath \"\"" Oct 01 10:40:45 crc kubenswrapper[4787]: I1001 10:40:45.715525 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x2bwg/crc-debug-2sbvd" event={"ID":"f4388fd5-b907-4c60-aa8f-e39f6a5ae2e4","Type":"ContainerDied","Data":"6c46935dfc876fbc6464b7ab220ff36dcd90e19dda873549257c553cc029fff0"} Oct 01 10:40:45 crc kubenswrapper[4787]: I1001 10:40:45.716135 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c46935dfc876fbc6464b7ab220ff36dcd90e19dda873549257c553cc029fff0" Oct 01 10:40:45 crc kubenswrapper[4787]: I1001 10:40:45.715633 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x2bwg/crc-debug-2sbvd" Oct 01 10:40:50 crc kubenswrapper[4787]: I1001 10:40:50.136955 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-x2bwg/crc-debug-2sbvd"] Oct 01 10:40:50 crc kubenswrapper[4787]: I1001 10:40:50.148572 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-x2bwg/crc-debug-2sbvd"] Oct 01 10:40:50 crc kubenswrapper[4787]: I1001 10:40:50.536477 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4388fd5-b907-4c60-aa8f-e39f6a5ae2e4" path="/var/lib/kubelet/pods/f4388fd5-b907-4c60-aa8f-e39f6a5ae2e4/volumes" Oct 01 10:40:51 crc kubenswrapper[4787]: I1001 10:40:51.327804 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-x2bwg/crc-debug-4b7hx"] Oct 01 10:40:51 crc kubenswrapper[4787]: E1001 10:40:51.328539 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4388fd5-b907-4c60-aa8f-e39f6a5ae2e4" containerName="container-00" Oct 01 10:40:51 crc kubenswrapper[4787]: I1001 10:40:51.328556 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4388fd5-b907-4c60-aa8f-e39f6a5ae2e4" containerName="container-00" Oct 01 10:40:51 crc kubenswrapper[4787]: I1001 10:40:51.328794 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4388fd5-b907-4c60-aa8f-e39f6a5ae2e4" containerName="container-00" Oct 01 10:40:51 crc kubenswrapper[4787]: I1001 10:40:51.329420 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x2bwg/crc-debug-4b7hx" Oct 01 10:40:51 crc kubenswrapper[4787]: I1001 10:40:51.331683 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-x2bwg"/"default-dockercfg-dnxsc" Oct 01 10:40:51 crc kubenswrapper[4787]: I1001 10:40:51.427667 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/809fa5fa-66c4-4b1b-8842-d3103a925bb2-host\") pod \"crc-debug-4b7hx\" (UID: \"809fa5fa-66c4-4b1b-8842-d3103a925bb2\") " pod="openshift-must-gather-x2bwg/crc-debug-4b7hx" Oct 01 10:40:51 crc kubenswrapper[4787]: I1001 10:40:51.427748 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vw8n8\" (UniqueName: \"kubernetes.io/projected/809fa5fa-66c4-4b1b-8842-d3103a925bb2-kube-api-access-vw8n8\") pod \"crc-debug-4b7hx\" (UID: \"809fa5fa-66c4-4b1b-8842-d3103a925bb2\") " pod="openshift-must-gather-x2bwg/crc-debug-4b7hx" Oct 01 10:40:51 crc kubenswrapper[4787]: I1001 10:40:51.534399 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/809fa5fa-66c4-4b1b-8842-d3103a925bb2-host\") pod \"crc-debug-4b7hx\" (UID: \"809fa5fa-66c4-4b1b-8842-d3103a925bb2\") " pod="openshift-must-gather-x2bwg/crc-debug-4b7hx" Oct 01 10:40:51 crc kubenswrapper[4787]: I1001 10:40:51.534490 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vw8n8\" (UniqueName: \"kubernetes.io/projected/809fa5fa-66c4-4b1b-8842-d3103a925bb2-kube-api-access-vw8n8\") pod \"crc-debug-4b7hx\" (UID: \"809fa5fa-66c4-4b1b-8842-d3103a925bb2\") " pod="openshift-must-gather-x2bwg/crc-debug-4b7hx" Oct 01 10:40:51 crc kubenswrapper[4787]: I1001 10:40:51.534852 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/809fa5fa-66c4-4b1b-8842-d3103a925bb2-host\") pod \"crc-debug-4b7hx\" (UID: \"809fa5fa-66c4-4b1b-8842-d3103a925bb2\") " pod="openshift-must-gather-x2bwg/crc-debug-4b7hx" Oct 01 10:40:51 crc kubenswrapper[4787]: I1001 10:40:51.562029 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vw8n8\" (UniqueName: \"kubernetes.io/projected/809fa5fa-66c4-4b1b-8842-d3103a925bb2-kube-api-access-vw8n8\") pod \"crc-debug-4b7hx\" (UID: \"809fa5fa-66c4-4b1b-8842-d3103a925bb2\") " pod="openshift-must-gather-x2bwg/crc-debug-4b7hx" Oct 01 10:40:51 crc kubenswrapper[4787]: I1001 10:40:51.648473 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x2bwg/crc-debug-4b7hx" Oct 01 10:40:51 crc kubenswrapper[4787]: I1001 10:40:51.779446 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x2bwg/crc-debug-4b7hx" event={"ID":"809fa5fa-66c4-4b1b-8842-d3103a925bb2","Type":"ContainerStarted","Data":"d003b09a840dd080234925ae15336d673dcbd5bb898d8835a0c2aa7159cd4c40"} Oct 01 10:40:52 crc kubenswrapper[4787]: I1001 10:40:52.794333 4787 generic.go:334] "Generic (PLEG): container finished" podID="809fa5fa-66c4-4b1b-8842-d3103a925bb2" containerID="f1292028166ffba3fda4873b1849dd3c783bd05723979e8e23bf644ba19df0a5" exitCode=0 Oct 01 10:40:52 crc kubenswrapper[4787]: I1001 10:40:52.794761 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x2bwg/crc-debug-4b7hx" event={"ID":"809fa5fa-66c4-4b1b-8842-d3103a925bb2","Type":"ContainerDied","Data":"f1292028166ffba3fda4873b1849dd3c783bd05723979e8e23bf644ba19df0a5"} Oct 01 10:40:52 crc kubenswrapper[4787]: I1001 10:40:52.844882 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-x2bwg/crc-debug-4b7hx"] Oct 01 10:40:52 crc kubenswrapper[4787]: I1001 10:40:52.852412 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-x2bwg/crc-debug-4b7hx"] Oct 01 10:40:53 crc kubenswrapper[4787]: I1001 10:40:53.911602 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x2bwg/crc-debug-4b7hx" Oct 01 10:40:54 crc kubenswrapper[4787]: I1001 10:40:54.083823 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vw8n8\" (UniqueName: \"kubernetes.io/projected/809fa5fa-66c4-4b1b-8842-d3103a925bb2-kube-api-access-vw8n8\") pod \"809fa5fa-66c4-4b1b-8842-d3103a925bb2\" (UID: \"809fa5fa-66c4-4b1b-8842-d3103a925bb2\") " Oct 01 10:40:54 crc kubenswrapper[4787]: I1001 10:40:54.084213 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/809fa5fa-66c4-4b1b-8842-d3103a925bb2-host\") pod \"809fa5fa-66c4-4b1b-8842-d3103a925bb2\" (UID: \"809fa5fa-66c4-4b1b-8842-d3103a925bb2\") " Oct 01 10:40:54 crc kubenswrapper[4787]: I1001 10:40:54.084303 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/809fa5fa-66c4-4b1b-8842-d3103a925bb2-host" (OuterVolumeSpecName: "host") pod "809fa5fa-66c4-4b1b-8842-d3103a925bb2" (UID: "809fa5fa-66c4-4b1b-8842-d3103a925bb2"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 01 10:40:54 crc kubenswrapper[4787]: I1001 10:40:54.084840 4787 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/809fa5fa-66c4-4b1b-8842-d3103a925bb2-host\") on node \"crc\" DevicePath \"\"" Oct 01 10:40:54 crc kubenswrapper[4787]: I1001 10:40:54.092836 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/809fa5fa-66c4-4b1b-8842-d3103a925bb2-kube-api-access-vw8n8" (OuterVolumeSpecName: "kube-api-access-vw8n8") pod "809fa5fa-66c4-4b1b-8842-d3103a925bb2" (UID: "809fa5fa-66c4-4b1b-8842-d3103a925bb2"). InnerVolumeSpecName "kube-api-access-vw8n8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:40:54 crc kubenswrapper[4787]: I1001 10:40:54.186265 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vw8n8\" (UniqueName: \"kubernetes.io/projected/809fa5fa-66c4-4b1b-8842-d3103a925bb2-kube-api-access-vw8n8\") on node \"crc\" DevicePath \"\"" Oct 01 10:40:54 crc kubenswrapper[4787]: I1001 10:40:54.507859 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh_a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1/util/0.log" Oct 01 10:40:54 crc kubenswrapper[4787]: I1001 10:40:54.536120 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="809fa5fa-66c4-4b1b-8842-d3103a925bb2" path="/var/lib/kubelet/pods/809fa5fa-66c4-4b1b-8842-d3103a925bb2/volumes" Oct 01 10:40:54 crc kubenswrapper[4787]: I1001 10:40:54.747001 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh_a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1/util/0.log" Oct 01 10:40:54 crc kubenswrapper[4787]: I1001 10:40:54.748246 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh_a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1/pull/0.log" Oct 01 10:40:54 crc kubenswrapper[4787]: I1001 10:40:54.761714 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh_a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1/pull/0.log" Oct 01 10:40:54 crc kubenswrapper[4787]: I1001 10:40:54.815991 4787 scope.go:117] "RemoveContainer" containerID="f1292028166ffba3fda4873b1849dd3c783bd05723979e8e23bf644ba19df0a5" Oct 01 10:40:54 crc kubenswrapper[4787]: I1001 10:40:54.816025 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x2bwg/crc-debug-4b7hx" Oct 01 10:40:54 crc kubenswrapper[4787]: I1001 10:40:54.959487 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh_a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1/util/0.log" Oct 01 10:40:54 crc kubenswrapper[4787]: I1001 10:40:54.996662 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh_a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1/pull/0.log" Oct 01 10:40:55 crc kubenswrapper[4787]: I1001 10:40:55.006285 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0e56acba034a4fe5e21c87878f7cabfd0ead2befdef111b141ca11c71amnpgh_a3d12d42-ce80-4a5c-8de6-ef1ef0aa8bb1/extract/0.log" Oct 01 10:40:55 crc kubenswrapper[4787]: I1001 10:40:55.147437 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-f7f98cb69-wqs6n_26b292ae-dcb2-41f6-86b5-2eddfecdf356/kube-rbac-proxy/0.log" Oct 01 10:40:55 crc kubenswrapper[4787]: I1001 10:40:55.215028 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-f7f98cb69-wqs6n_26b292ae-dcb2-41f6-86b5-2eddfecdf356/manager/0.log" Oct 01 10:40:55 crc kubenswrapper[4787]: I1001 10:40:55.239897 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859cd486d-6h8cm_24354b04-0daa-4fd4-8d38-b21f64bc2cfe/kube-rbac-proxy/0.log" Oct 01 10:40:55 crc kubenswrapper[4787]: I1001 10:40:55.395411 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859cd486d-6h8cm_24354b04-0daa-4fd4-8d38-b21f64bc2cfe/manager/0.log" Oct 01 10:40:55 crc kubenswrapper[4787]: I1001 10:40:55.487133 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-77fb7bcf5b-2pxzs_77e07ae2-1dae-4570-84b8-c137e86ac628/kube-rbac-proxy/0.log" Oct 01 10:40:55 crc kubenswrapper[4787]: I1001 10:40:55.495555 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-77fb7bcf5b-2pxzs_77e07ae2-1dae-4570-84b8-c137e86ac628/manager/0.log" Oct 01 10:40:55 crc kubenswrapper[4787]: I1001 10:40:55.638033 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-8bc4775b5-x2v89_b4902328-fb90-40f9-a84c-5a0359d699d4/kube-rbac-proxy/0.log" Oct 01 10:40:55 crc kubenswrapper[4787]: I1001 10:40:55.792273 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-8bc4775b5-x2v89_b4902328-fb90-40f9-a84c-5a0359d699d4/manager/0.log" Oct 01 10:40:55 crc kubenswrapper[4787]: I1001 10:40:55.859897 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b4fc86755-lbbxd_b1a0ae18-3c6d-473a-a660-5ff59f6a87bf/kube-rbac-proxy/0.log" Oct 01 10:40:55 crc kubenswrapper[4787]: I1001 10:40:55.866145 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b4fc86755-lbbxd_b1a0ae18-3c6d-473a-a660-5ff59f6a87bf/manager/0.log" Oct 01 10:40:56 crc kubenswrapper[4787]: I1001 10:40:56.012315 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-679b4759bb-qt9bq_8f52292a-2603-4524-b3fc-8d7b003f95f8/kube-rbac-proxy/0.log" Oct 01 10:40:56 crc kubenswrapper[4787]: I1001 10:40:56.071062 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-679b4759bb-qt9bq_8f52292a-2603-4524-b3fc-8d7b003f95f8/manager/0.log" Oct 01 10:40:56 crc kubenswrapper[4787]: I1001 10:40:56.162147 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5c8fdc4d5c-t72zp_9e03d54f-d462-4894-9c3b-8253a861af54/kube-rbac-proxy/0.log" Oct 01 10:40:56 crc kubenswrapper[4787]: I1001 10:40:56.367447 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5f45cd594f-zrtv2_ae2ba2f0-8422-4a03-b64f-d335f76c5532/kube-rbac-proxy/0.log" Oct 01 10:40:56 crc kubenswrapper[4787]: I1001 10:40:56.402143 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5c8fdc4d5c-t72zp_9e03d54f-d462-4894-9c3b-8253a861af54/manager/0.log" Oct 01 10:40:56 crc kubenswrapper[4787]: I1001 10:40:56.427138 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5f45cd594f-zrtv2_ae2ba2f0-8422-4a03-b64f-d335f76c5532/manager/0.log" Oct 01 10:40:56 crc kubenswrapper[4787]: I1001 10:40:56.590518 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-59d7dc95cf-9tdp2_88d93cbc-27c2-47be-a882-6607925d4391/kube-rbac-proxy/0.log" Oct 01 10:40:56 crc kubenswrapper[4787]: I1001 10:40:56.669464 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-59d7dc95cf-9tdp2_88d93cbc-27c2-47be-a882-6607925d4391/manager/0.log" Oct 01 10:40:56 crc kubenswrapper[4787]: I1001 10:40:56.801722 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-b7cf8cb5f-5qz2m_2dfc2636-198f-4b51-b07e-40ff76c1091c/manager/0.log" Oct 01 10:40:56 crc kubenswrapper[4787]: I1001 10:40:56.811069 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-b7cf8cb5f-5qz2m_2dfc2636-198f-4b51-b07e-40ff76c1091c/kube-rbac-proxy/0.log" Oct 01 10:40:56 crc kubenswrapper[4787]: I1001 10:40:56.901390 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-67bf5bb885-qmr4v_b54a691a-cd96-4ae4-a9ce-65662f58cd60/kube-rbac-proxy/0.log" Oct 01 10:40:57 crc kubenswrapper[4787]: I1001 10:40:57.052131 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-67bf5bb885-qmr4v_b54a691a-cd96-4ae4-a9ce-65662f58cd60/manager/0.log" Oct 01 10:40:57 crc kubenswrapper[4787]: I1001 10:40:57.080350 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-54fbbfcd44-4jlcz_8b59a5e1-4f45-4bcd-b9ce-63bae8b5ed57/kube-rbac-proxy/0.log" Oct 01 10:40:57 crc kubenswrapper[4787]: I1001 10:40:57.177926 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-54fbbfcd44-4jlcz_8b59a5e1-4f45-4bcd-b9ce-63bae8b5ed57/manager/0.log" Oct 01 10:40:57 crc kubenswrapper[4787]: I1001 10:40:57.311634 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7fd5b6bbc6-nbxwx_c186a6f1-61b9-4ab7-8996-5e1d8439a77c/kube-rbac-proxy/0.log" Oct 01 10:40:57 crc kubenswrapper[4787]: I1001 10:40:57.404279 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7fd5b6bbc6-nbxwx_c186a6f1-61b9-4ab7-8996-5e1d8439a77c/manager/0.log" Oct 01 10:40:57 crc kubenswrapper[4787]: I1001 10:40:57.523120 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-75f8d67d86-xjbwt_20e0b9a6-f51b-4925-bffd-8fa00e9a8ec7/manager/0.log" Oct 01 10:40:57 crc kubenswrapper[4787]: I1001 10:40:57.528971 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-75f8d67d86-xjbwt_20e0b9a6-f51b-4925-bffd-8fa00e9a8ec7/kube-rbac-proxy/0.log" Oct 01 10:40:57 crc kubenswrapper[4787]: I1001 10:40:57.638731 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-659bb84579xswrt_2d5ffd79-67fa-4dd0-81a7-b44156946b86/kube-rbac-proxy/0.log" Oct 01 10:40:57 crc kubenswrapper[4787]: I1001 10:40:57.689896 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-659bb84579xswrt_2d5ffd79-67fa-4dd0-81a7-b44156946b86/manager/0.log" Oct 01 10:40:57 crc kubenswrapper[4787]: I1001 10:40:57.750200 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6c7b6bcb7c-bxglr_40d23171-497b-47d9-96f4-51f32d7683a1/kube-rbac-proxy/0.log" Oct 01 10:40:57 crc kubenswrapper[4787]: I1001 10:40:57.911118 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-676c66f88b-7hgdc_faaa86ca-d16f-4ee1-9a9a-15c54aca2a8b/kube-rbac-proxy/0.log" Oct 01 10:40:58 crc kubenswrapper[4787]: I1001 10:40:58.224824 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-676c66f88b-7hgdc_faaa86ca-d16f-4ee1-9a9a-15c54aca2a8b/operator/0.log" Oct 01 10:40:58 crc kubenswrapper[4787]: I1001 10:40:58.231505 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-jnbw4_c6b4eeb7-99c2-44ff-9f71-e27c17c17649/registry-server/0.log" Oct 01 10:40:58 crc kubenswrapper[4787]: I1001 10:40:58.269159 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-84c745747f-8qtcw_57f9bf14-0074-43ba-882d-a9102b047a3b/kube-rbac-proxy/0.log" Oct 01 10:40:58 crc kubenswrapper[4787]: I1001 10:40:58.469868 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-84c745747f-8qtcw_57f9bf14-0074-43ba-882d-a9102b047a3b/manager/0.log" Oct 01 10:40:58 crc kubenswrapper[4787]: I1001 10:40:58.504901 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-598c4c8547-4zm2q_5024aa9b-39fc-4e58-96fe-c66d054e3d5a/kube-rbac-proxy/0.log" Oct 01 10:40:58 crc kubenswrapper[4787]: I1001 10:40:58.605734 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-598c4c8547-4zm2q_5024aa9b-39fc-4e58-96fe-c66d054e3d5a/manager/0.log" Oct 01 10:40:58 crc kubenswrapper[4787]: I1001 10:40:58.821463 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-xvd4r_f31cb2c2-eabc-4bca-88c5-2aa194fb13b5/operator/0.log" Oct 01 10:40:58 crc kubenswrapper[4787]: I1001 10:40:58.897755 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-689b4f76c9-xvrsn_ed13cb34-e846-41ab-aff7-8e5d4acd2e6b/kube-rbac-proxy/0.log" Oct 01 10:40:58 crc kubenswrapper[4787]: I1001 10:40:58.902616 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6c7b6bcb7c-bxglr_40d23171-497b-47d9-96f4-51f32d7683a1/manager/0.log" Oct 01 10:40:58 crc kubenswrapper[4787]: I1001 10:40:58.991320 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-689b4f76c9-xvrsn_ed13cb34-e846-41ab-aff7-8e5d4acd2e6b/manager/0.log" Oct 01 10:40:59 crc kubenswrapper[4787]: I1001 10:40:59.067313 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-cb66d6b59-fqgnk_7139ac70-d405-449b-924e-01fd94e1375a/kube-rbac-proxy/0.log" Oct 01 10:40:59 crc kubenswrapper[4787]: I1001 10:40:59.133065 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-cb66d6b59-fqgnk_7139ac70-d405-449b-924e-01fd94e1375a/manager/0.log" Oct 01 10:40:59 crc kubenswrapper[4787]: I1001 10:40:59.203934 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-cbdf6dc66-z2gzf_a74d851b-a805-4138-9bf4-1a4546a2f30a/kube-rbac-proxy/0.log" Oct 01 10:40:59 crc kubenswrapper[4787]: I1001 10:40:59.238601 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-cbdf6dc66-z2gzf_a74d851b-a805-4138-9bf4-1a4546a2f30a/manager/0.log" Oct 01 10:40:59 crc kubenswrapper[4787]: I1001 10:40:59.331680 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-68d7bc5569-4mgdw_fb776e43-5f95-44f0-96c0-945b2681d8c1/kube-rbac-proxy/0.log" Oct 01 10:40:59 crc kubenswrapper[4787]: I1001 10:40:59.333037 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-68d7bc5569-4mgdw_fb776e43-5f95-44f0-96c0-945b2681d8c1/manager/0.log" Oct 01 10:41:15 crc kubenswrapper[4787]: I1001 10:41:15.582478 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-zssrz_e6b988c0-936d-4b49-80a2-015ae390f262/control-plane-machine-set-operator/0.log" Oct 01 10:41:15 crc kubenswrapper[4787]: I1001 10:41:15.748200 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-4g5xc_07892f22-556b-49a6-8f71-3d3b16b2cdef/machine-api-operator/0.log" Oct 01 10:41:15 crc kubenswrapper[4787]: I1001 10:41:15.777933 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-4g5xc_07892f22-556b-49a6-8f71-3d3b16b2cdef/kube-rbac-proxy/0.log" Oct 01 10:41:30 crc kubenswrapper[4787]: I1001 10:41:30.116032 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-w2pvt_b1f30c7b-06ed-4ccf-a3d0-8df65a3a8c67/cert-manager-controller/0.log" Oct 01 10:41:30 crc kubenswrapper[4787]: I1001 10:41:30.395307 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-ss8nf_081d3606-6bc1-4f9b-8f3b-6ee00ed9d164/cert-manager-cainjector/0.log" Oct 01 10:41:30 crc kubenswrapper[4787]: I1001 10:41:30.417816 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-xgrnp_426ff9a0-bcf0-486e-bd80-2d4de51be0cf/cert-manager-webhook/0.log" Oct 01 10:41:42 crc kubenswrapper[4787]: I1001 10:41:42.701928 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-864bb6dfb5-cjhqm_e4a805b1-d488-4007-9dd3-38d3fbab1f10/nmstate-console-plugin/0.log" Oct 01 10:41:42 crc kubenswrapper[4787]: I1001 10:41:42.958524 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-2rzr2_ee9e051d-7391-4a43-9a02-5084da577b36/nmstate-handler/0.log" Oct 01 10:41:43 crc kubenswrapper[4787]: I1001 10:41:43.053334 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58fcddf996-2cnst_10cd7a64-357c-4c0d-b11c-426e4060b61c/kube-rbac-proxy/0.log" Oct 01 10:41:43 crc kubenswrapper[4787]: I1001 10:41:43.055624 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58fcddf996-2cnst_10cd7a64-357c-4c0d-b11c-426e4060b61c/nmstate-metrics/0.log" Oct 01 10:41:43 crc kubenswrapper[4787]: I1001 10:41:43.283497 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5d6f6cfd66-8l5nr_27ab4c98-35a2-45be-864e-4550be0fa875/nmstate-operator/0.log" Oct 01 10:41:43 crc kubenswrapper[4787]: I1001 10:41:43.337164 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6d689559c5-22sp6_97efd9a6-e221-4aac-97b1-b49de45edd88/nmstate-webhook/0.log" Oct 01 10:41:57 crc kubenswrapper[4787]: I1001 10:41:57.715797 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5d688f5ffc-gnclz_9a498324-bbec-40e1-9649-f23ca26561f3/kube-rbac-proxy/0.log" Oct 01 10:41:57 crc kubenswrapper[4787]: I1001 10:41:57.801463 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5d688f5ffc-gnclz_9a498324-bbec-40e1-9649-f23ca26561f3/controller/0.log" Oct 01 10:41:57 crc kubenswrapper[4787]: I1001 10:41:57.950973 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-csxhg_c6570c99-4cff-4194-9744-0efc00f88332/cp-frr-files/0.log" Oct 01 10:41:58 crc kubenswrapper[4787]: I1001 10:41:58.135656 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-csxhg_c6570c99-4cff-4194-9744-0efc00f88332/cp-frr-files/0.log" Oct 01 10:41:58 crc kubenswrapper[4787]: I1001 10:41:58.158566 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-csxhg_c6570c99-4cff-4194-9744-0efc00f88332/cp-reloader/0.log" Oct 01 10:41:58 crc kubenswrapper[4787]: I1001 10:41:58.166096 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-csxhg_c6570c99-4cff-4194-9744-0efc00f88332/cp-reloader/0.log" Oct 01 10:41:58 crc kubenswrapper[4787]: I1001 10:41:58.218707 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-csxhg_c6570c99-4cff-4194-9744-0efc00f88332/cp-metrics/0.log" Oct 01 10:41:58 crc kubenswrapper[4787]: I1001 10:41:58.416325 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-csxhg_c6570c99-4cff-4194-9744-0efc00f88332/cp-metrics/0.log" Oct 01 10:41:58 crc kubenswrapper[4787]: I1001 10:41:58.439477 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-csxhg_c6570c99-4cff-4194-9744-0efc00f88332/cp-metrics/0.log" Oct 01 10:41:58 crc kubenswrapper[4787]: I1001 10:41:58.442709 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-csxhg_c6570c99-4cff-4194-9744-0efc00f88332/cp-frr-files/0.log" Oct 01 10:41:58 crc kubenswrapper[4787]: I1001 10:41:58.444876 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-csxhg_c6570c99-4cff-4194-9744-0efc00f88332/cp-reloader/0.log" Oct 01 10:41:58 crc kubenswrapper[4787]: I1001 10:41:58.638521 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-csxhg_c6570c99-4cff-4194-9744-0efc00f88332/cp-metrics/0.log" Oct 01 10:41:58 crc kubenswrapper[4787]: I1001 10:41:58.646937 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-csxhg_c6570c99-4cff-4194-9744-0efc00f88332/cp-reloader/0.log" Oct 01 10:41:58 crc kubenswrapper[4787]: I1001 10:41:58.661244 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-csxhg_c6570c99-4cff-4194-9744-0efc00f88332/cp-frr-files/0.log" Oct 01 10:41:58 crc kubenswrapper[4787]: I1001 10:41:58.715939 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-csxhg_c6570c99-4cff-4194-9744-0efc00f88332/controller/0.log" Oct 01 10:41:58 crc kubenswrapper[4787]: I1001 10:41:58.879944 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-csxhg_c6570c99-4cff-4194-9744-0efc00f88332/frr-metrics/0.log" Oct 01 10:41:58 crc kubenswrapper[4787]: I1001 10:41:58.927932 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-csxhg_c6570c99-4cff-4194-9744-0efc00f88332/kube-rbac-proxy/0.log" Oct 01 10:41:58 crc kubenswrapper[4787]: I1001 10:41:58.990133 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-csxhg_c6570c99-4cff-4194-9744-0efc00f88332/kube-rbac-proxy-frr/0.log" Oct 01 10:41:59 crc kubenswrapper[4787]: I1001 10:41:59.155015 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-csxhg_c6570c99-4cff-4194-9744-0efc00f88332/reloader/0.log" Oct 01 10:41:59 crc kubenswrapper[4787]: I1001 10:41:59.227118 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-5478bdb765-kcxb8_8b191fd2-3066-4e07-9641-98e44be362f5/frr-k8s-webhook-server/0.log" Oct 01 10:41:59 crc kubenswrapper[4787]: I1001 10:41:59.442508 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-86f7dcd4cf-c4xjr_178c8464-8b87-4dc1-9c52-7dc1f5b3714d/manager/0.log" Oct 01 10:41:59 crc kubenswrapper[4787]: I1001 10:41:59.602184 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-7d48896987-v5r98_55c631b4-c7e0-47cb-9e66-93b8a8c51877/webhook-server/0.log" Oct 01 10:41:59 crc kubenswrapper[4787]: I1001 10:41:59.852359 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-cc45k_c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668/kube-rbac-proxy/0.log" Oct 01 10:42:00 crc kubenswrapper[4787]: I1001 10:42:00.455476 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-cc45k_c9ad2dd2-b456-4e2c-a1d0-5c7b8664c668/speaker/0.log" Oct 01 10:42:00 crc kubenswrapper[4787]: I1001 10:42:00.696903 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-csxhg_c6570c99-4cff-4194-9744-0efc00f88332/frr/0.log" Oct 01 10:42:12 crc kubenswrapper[4787]: I1001 10:42:12.537892 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc_e60e8301-91e9-45d5-a3d1-cd29ce1e9514/util/0.log" Oct 01 10:42:12 crc kubenswrapper[4787]: I1001 10:42:12.738380 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc_e60e8301-91e9-45d5-a3d1-cd29ce1e9514/util/0.log" Oct 01 10:42:12 crc kubenswrapper[4787]: I1001 10:42:12.781311 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc_e60e8301-91e9-45d5-a3d1-cd29ce1e9514/pull/0.log" Oct 01 10:42:12 crc kubenswrapper[4787]: I1001 10:42:12.801190 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc_e60e8301-91e9-45d5-a3d1-cd29ce1e9514/pull/0.log" Oct 01 10:42:12 crc kubenswrapper[4787]: I1001 10:42:12.982273 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc_e60e8301-91e9-45d5-a3d1-cd29ce1e9514/util/0.log" Oct 01 10:42:13 crc kubenswrapper[4787]: I1001 10:42:13.007801 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc_e60e8301-91e9-45d5-a3d1-cd29ce1e9514/extract/0.log" Oct 01 10:42:13 crc kubenswrapper[4787]: I1001 10:42:13.008490 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bczbzxc_e60e8301-91e9-45d5-a3d1-cd29ce1e9514/pull/0.log" Oct 01 10:42:13 crc kubenswrapper[4787]: I1001 10:42:13.192163 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8ksjp_09d257f5-ca13-42c9-92f8-920f99ebadc1/extract-utilities/0.log" Oct 01 10:42:13 crc kubenswrapper[4787]: I1001 10:42:13.352584 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8ksjp_09d257f5-ca13-42c9-92f8-920f99ebadc1/extract-utilities/0.log" Oct 01 10:42:13 crc kubenswrapper[4787]: I1001 10:42:13.391561 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8ksjp_09d257f5-ca13-42c9-92f8-920f99ebadc1/extract-content/0.log" Oct 01 10:42:13 crc kubenswrapper[4787]: I1001 10:42:13.405646 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8ksjp_09d257f5-ca13-42c9-92f8-920f99ebadc1/extract-content/0.log" Oct 01 10:42:13 crc kubenswrapper[4787]: I1001 10:42:13.639940 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8ksjp_09d257f5-ca13-42c9-92f8-920f99ebadc1/extract-content/0.log" Oct 01 10:42:13 crc kubenswrapper[4787]: I1001 10:42:13.705059 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8ksjp_09d257f5-ca13-42c9-92f8-920f99ebadc1/extract-utilities/0.log" Oct 01 10:42:13 crc kubenswrapper[4787]: I1001 10:42:13.866726 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4m4pm_6e4be47d-87d5-469d-9ad2-b12efb75e1aa/extract-utilities/0.log" Oct 01 10:42:14 crc kubenswrapper[4787]: I1001 10:42:14.170738 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8ksjp_09d257f5-ca13-42c9-92f8-920f99ebadc1/registry-server/0.log" Oct 01 10:42:14 crc kubenswrapper[4787]: I1001 10:42:14.219148 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4m4pm_6e4be47d-87d5-469d-9ad2-b12efb75e1aa/extract-utilities/0.log" Oct 01 10:42:14 crc kubenswrapper[4787]: I1001 10:42:14.236924 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4m4pm_6e4be47d-87d5-469d-9ad2-b12efb75e1aa/extract-content/0.log" Oct 01 10:42:14 crc kubenswrapper[4787]: I1001 10:42:14.250222 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4m4pm_6e4be47d-87d5-469d-9ad2-b12efb75e1aa/extract-content/0.log" Oct 01 10:42:14 crc kubenswrapper[4787]: I1001 10:42:14.384461 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4m4pm_6e4be47d-87d5-469d-9ad2-b12efb75e1aa/extract-utilities/0.log" Oct 01 10:42:14 crc kubenswrapper[4787]: I1001 10:42:14.461760 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4m4pm_6e4be47d-87d5-469d-9ad2-b12efb75e1aa/extract-content/0.log" Oct 01 10:42:14 crc kubenswrapper[4787]: I1001 10:42:14.642967 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft_810c9efa-ff37-4f0a-8c30-d620a00bd8a4/util/0.log" Oct 01 10:42:14 crc kubenswrapper[4787]: I1001 10:42:14.868397 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft_810c9efa-ff37-4f0a-8c30-d620a00bd8a4/util/0.log" Oct 01 10:42:14 crc kubenswrapper[4787]: I1001 10:42:14.922718 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft_810c9efa-ff37-4f0a-8c30-d620a00bd8a4/pull/0.log" Oct 01 10:42:14 crc kubenswrapper[4787]: I1001 10:42:14.927674 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft_810c9efa-ff37-4f0a-8c30-d620a00bd8a4/pull/0.log" Oct 01 10:42:15 crc kubenswrapper[4787]: I1001 10:42:15.060712 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4m4pm_6e4be47d-87d5-469d-9ad2-b12efb75e1aa/registry-server/0.log" Oct 01 10:42:15 crc kubenswrapper[4787]: I1001 10:42:15.188851 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft_810c9efa-ff37-4f0a-8c30-d620a00bd8a4/pull/0.log" Oct 01 10:42:15 crc kubenswrapper[4787]: I1001 10:42:15.192294 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft_810c9efa-ff37-4f0a-8c30-d620a00bd8a4/extract/0.log" Oct 01 10:42:15 crc kubenswrapper[4787]: I1001 10:42:15.201749 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d967drft_810c9efa-ff37-4f0a-8c30-d620a00bd8a4/util/0.log" Oct 01 10:42:16 crc kubenswrapper[4787]: I1001 10:42:16.228438 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-6zb6m_3390abc0-6474-4a34-8e95-c5923f699e80/marketplace-operator/0.log" Oct 01 10:42:16 crc kubenswrapper[4787]: I1001 10:42:16.258691 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-swbhm_593f4070-9419-4e2f-9da0-741e86877588/extract-utilities/0.log" Oct 01 10:42:16 crc kubenswrapper[4787]: I1001 10:42:16.497917 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-swbhm_593f4070-9419-4e2f-9da0-741e86877588/extract-content/0.log" Oct 01 10:42:16 crc kubenswrapper[4787]: I1001 10:42:16.513675 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-swbhm_593f4070-9419-4e2f-9da0-741e86877588/extract-utilities/0.log" Oct 01 10:42:16 crc kubenswrapper[4787]: I1001 10:42:16.530048 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-swbhm_593f4070-9419-4e2f-9da0-741e86877588/extract-content/0.log" Oct 01 10:42:16 crc kubenswrapper[4787]: I1001 10:42:16.730302 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-swbhm_593f4070-9419-4e2f-9da0-741e86877588/extract-content/0.log" Oct 01 10:42:16 crc kubenswrapper[4787]: I1001 10:42:16.742565 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-swbhm_593f4070-9419-4e2f-9da0-741e86877588/extract-utilities/0.log" Oct 01 10:42:16 crc kubenswrapper[4787]: I1001 10:42:16.838019 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pwj4z_c3868e91-d922-4768-af6a-123795df5a28/extract-utilities/0.log" Oct 01 10:42:16 crc kubenswrapper[4787]: I1001 10:42:16.876694 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-swbhm_593f4070-9419-4e2f-9da0-741e86877588/registry-server/0.log" Oct 01 10:42:17 crc kubenswrapper[4787]: I1001 10:42:17.080357 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pwj4z_c3868e91-d922-4768-af6a-123795df5a28/extract-content/0.log" Oct 01 10:42:17 crc kubenswrapper[4787]: I1001 10:42:17.101487 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pwj4z_c3868e91-d922-4768-af6a-123795df5a28/extract-content/0.log" Oct 01 10:42:17 crc kubenswrapper[4787]: I1001 10:42:17.102007 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pwj4z_c3868e91-d922-4768-af6a-123795df5a28/extract-utilities/0.log" Oct 01 10:42:17 crc kubenswrapper[4787]: I1001 10:42:17.284022 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pwj4z_c3868e91-d922-4768-af6a-123795df5a28/extract-content/0.log" Oct 01 10:42:17 crc kubenswrapper[4787]: I1001 10:42:17.296283 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pwj4z_c3868e91-d922-4768-af6a-123795df5a28/extract-utilities/0.log" Oct 01 10:42:17 crc kubenswrapper[4787]: I1001 10:42:17.773608 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pwj4z_c3868e91-d922-4768-af6a-123795df5a28/registry-server/0.log" Oct 01 10:42:41 crc kubenswrapper[4787]: I1001 10:42:41.250555 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:42:41 crc kubenswrapper[4787]: I1001 10:42:41.252126 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:42:57 crc kubenswrapper[4787]: I1001 10:42:57.311303 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xmvqc"] Oct 01 10:42:57 crc kubenswrapper[4787]: E1001 10:42:57.312302 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="809fa5fa-66c4-4b1b-8842-d3103a925bb2" containerName="container-00" Oct 01 10:42:57 crc kubenswrapper[4787]: I1001 10:42:57.312318 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="809fa5fa-66c4-4b1b-8842-d3103a925bb2" containerName="container-00" Oct 01 10:42:57 crc kubenswrapper[4787]: I1001 10:42:57.312504 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="809fa5fa-66c4-4b1b-8842-d3103a925bb2" containerName="container-00" Oct 01 10:42:57 crc kubenswrapper[4787]: I1001 10:42:57.313906 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xmvqc" Oct 01 10:42:57 crc kubenswrapper[4787]: I1001 10:42:57.321572 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xmvqc"] Oct 01 10:42:57 crc kubenswrapper[4787]: I1001 10:42:57.428790 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9f49564-58da-4338-9bff-a9c5b64b78df-utilities\") pod \"redhat-operators-xmvqc\" (UID: \"c9f49564-58da-4338-9bff-a9c5b64b78df\") " pod="openshift-marketplace/redhat-operators-xmvqc" Oct 01 10:42:57 crc kubenswrapper[4787]: I1001 10:42:57.428898 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c56lf\" (UniqueName: \"kubernetes.io/projected/c9f49564-58da-4338-9bff-a9c5b64b78df-kube-api-access-c56lf\") pod \"redhat-operators-xmvqc\" (UID: \"c9f49564-58da-4338-9bff-a9c5b64b78df\") " pod="openshift-marketplace/redhat-operators-xmvqc" Oct 01 10:42:57 crc kubenswrapper[4787]: I1001 10:42:57.429054 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9f49564-58da-4338-9bff-a9c5b64b78df-catalog-content\") pod \"redhat-operators-xmvqc\" (UID: \"c9f49564-58da-4338-9bff-a9c5b64b78df\") " pod="openshift-marketplace/redhat-operators-xmvqc" Oct 01 10:42:57 crc kubenswrapper[4787]: I1001 10:42:57.530330 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9f49564-58da-4338-9bff-a9c5b64b78df-utilities\") pod \"redhat-operators-xmvqc\" (UID: \"c9f49564-58da-4338-9bff-a9c5b64b78df\") " pod="openshift-marketplace/redhat-operators-xmvqc" Oct 01 10:42:57 crc kubenswrapper[4787]: I1001 10:42:57.530373 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c56lf\" (UniqueName: \"kubernetes.io/projected/c9f49564-58da-4338-9bff-a9c5b64b78df-kube-api-access-c56lf\") pod \"redhat-operators-xmvqc\" (UID: \"c9f49564-58da-4338-9bff-a9c5b64b78df\") " pod="openshift-marketplace/redhat-operators-xmvqc" Oct 01 10:42:57 crc kubenswrapper[4787]: I1001 10:42:57.530435 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9f49564-58da-4338-9bff-a9c5b64b78df-catalog-content\") pod \"redhat-operators-xmvqc\" (UID: \"c9f49564-58da-4338-9bff-a9c5b64b78df\") " pod="openshift-marketplace/redhat-operators-xmvqc" Oct 01 10:42:57 crc kubenswrapper[4787]: I1001 10:42:57.530925 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9f49564-58da-4338-9bff-a9c5b64b78df-utilities\") pod \"redhat-operators-xmvqc\" (UID: \"c9f49564-58da-4338-9bff-a9c5b64b78df\") " pod="openshift-marketplace/redhat-operators-xmvqc" Oct 01 10:42:57 crc kubenswrapper[4787]: I1001 10:42:57.530952 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9f49564-58da-4338-9bff-a9c5b64b78df-catalog-content\") pod \"redhat-operators-xmvqc\" (UID: \"c9f49564-58da-4338-9bff-a9c5b64b78df\") " pod="openshift-marketplace/redhat-operators-xmvqc" Oct 01 10:42:57 crc kubenswrapper[4787]: I1001 10:42:57.550405 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c56lf\" (UniqueName: \"kubernetes.io/projected/c9f49564-58da-4338-9bff-a9c5b64b78df-kube-api-access-c56lf\") pod \"redhat-operators-xmvqc\" (UID: \"c9f49564-58da-4338-9bff-a9c5b64b78df\") " pod="openshift-marketplace/redhat-operators-xmvqc" Oct 01 10:42:57 crc kubenswrapper[4787]: I1001 10:42:57.638514 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xmvqc" Oct 01 10:42:58 crc kubenswrapper[4787]: I1001 10:42:58.700882 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xmvqc"] Oct 01 10:42:59 crc kubenswrapper[4787]: I1001 10:42:59.052668 4787 generic.go:334] "Generic (PLEG): container finished" podID="c9f49564-58da-4338-9bff-a9c5b64b78df" containerID="444d93b283bb8179cf67353e6731cddca7b5bbd1ce486a6d65bf38a73186f0f0" exitCode=0 Oct 01 10:42:59 crc kubenswrapper[4787]: I1001 10:42:59.052724 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xmvqc" event={"ID":"c9f49564-58da-4338-9bff-a9c5b64b78df","Type":"ContainerDied","Data":"444d93b283bb8179cf67353e6731cddca7b5bbd1ce486a6d65bf38a73186f0f0"} Oct 01 10:42:59 crc kubenswrapper[4787]: I1001 10:42:59.052933 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xmvqc" event={"ID":"c9f49564-58da-4338-9bff-a9c5b64b78df","Type":"ContainerStarted","Data":"008a8882b6eeb6192a493c576cbfcf82bdefcf7d216678c82890a5787c02304c"} Oct 01 10:42:59 crc kubenswrapper[4787]: I1001 10:42:59.054744 4787 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 01 10:43:01 crc kubenswrapper[4787]: I1001 10:43:01.080677 4787 generic.go:334] "Generic (PLEG): container finished" podID="c9f49564-58da-4338-9bff-a9c5b64b78df" containerID="8a9a1ff5393a4b3c1f49ff34eac342bc2669c127fdddd5d6dbbccedb81175ecb" exitCode=0 Oct 01 10:43:01 crc kubenswrapper[4787]: I1001 10:43:01.080721 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xmvqc" event={"ID":"c9f49564-58da-4338-9bff-a9c5b64b78df","Type":"ContainerDied","Data":"8a9a1ff5393a4b3c1f49ff34eac342bc2669c127fdddd5d6dbbccedb81175ecb"} Oct 01 10:43:01 crc kubenswrapper[4787]: I1001 10:43:01.910105 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-f2m6d"] Oct 01 10:43:01 crc kubenswrapper[4787]: I1001 10:43:01.912549 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f2m6d" Oct 01 10:43:01 crc kubenswrapper[4787]: I1001 10:43:01.922571 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f2m6d"] Oct 01 10:43:02 crc kubenswrapper[4787]: I1001 10:43:02.027217 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e297954-d108-4f85-ab10-02634bd69bd3-utilities\") pod \"community-operators-f2m6d\" (UID: \"0e297954-d108-4f85-ab10-02634bd69bd3\") " pod="openshift-marketplace/community-operators-f2m6d" Oct 01 10:43:02 crc kubenswrapper[4787]: I1001 10:43:02.027293 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmh2x\" (UniqueName: \"kubernetes.io/projected/0e297954-d108-4f85-ab10-02634bd69bd3-kube-api-access-wmh2x\") pod \"community-operators-f2m6d\" (UID: \"0e297954-d108-4f85-ab10-02634bd69bd3\") " pod="openshift-marketplace/community-operators-f2m6d" Oct 01 10:43:02 crc kubenswrapper[4787]: I1001 10:43:02.027365 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e297954-d108-4f85-ab10-02634bd69bd3-catalog-content\") pod \"community-operators-f2m6d\" (UID: \"0e297954-d108-4f85-ab10-02634bd69bd3\") " pod="openshift-marketplace/community-operators-f2m6d" Oct 01 10:43:02 crc kubenswrapper[4787]: I1001 10:43:02.092972 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xmvqc" event={"ID":"c9f49564-58da-4338-9bff-a9c5b64b78df","Type":"ContainerStarted","Data":"6fdb34304eccf388d9ce6c1fcab86ad9ba44995d1fd22e017a3b3c21e6b48291"} Oct 01 10:43:02 crc kubenswrapper[4787]: I1001 10:43:02.110959 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kftxb"] Oct 01 10:43:02 crc kubenswrapper[4787]: I1001 10:43:02.113451 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kftxb" Oct 01 10:43:02 crc kubenswrapper[4787]: I1001 10:43:02.121005 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kftxb"] Oct 01 10:43:02 crc kubenswrapper[4787]: I1001 10:43:02.127596 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xmvqc" podStartSLOduration=2.619218601 podStartE2EDuration="5.127579212s" podCreationTimestamp="2025-10-01 10:42:57 +0000 UTC" firstStartedPulling="2025-10-01 10:42:59.054380491 +0000 UTC m=+4011.169524658" lastFinishedPulling="2025-10-01 10:43:01.562741092 +0000 UTC m=+4013.677885269" observedRunningTime="2025-10-01 10:43:02.119379671 +0000 UTC m=+4014.234523828" watchObservedRunningTime="2025-10-01 10:43:02.127579212 +0000 UTC m=+4014.242723369" Oct 01 10:43:02 crc kubenswrapper[4787]: I1001 10:43:02.129313 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e297954-d108-4f85-ab10-02634bd69bd3-utilities\") pod \"community-operators-f2m6d\" (UID: \"0e297954-d108-4f85-ab10-02634bd69bd3\") " pod="openshift-marketplace/community-operators-f2m6d" Oct 01 10:43:02 crc kubenswrapper[4787]: I1001 10:43:02.129406 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmh2x\" (UniqueName: \"kubernetes.io/projected/0e297954-d108-4f85-ab10-02634bd69bd3-kube-api-access-wmh2x\") pod \"community-operators-f2m6d\" (UID: \"0e297954-d108-4f85-ab10-02634bd69bd3\") " pod="openshift-marketplace/community-operators-f2m6d" Oct 01 10:43:02 crc kubenswrapper[4787]: I1001 10:43:02.129465 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e297954-d108-4f85-ab10-02634bd69bd3-catalog-content\") pod \"community-operators-f2m6d\" (UID: \"0e297954-d108-4f85-ab10-02634bd69bd3\") " pod="openshift-marketplace/community-operators-f2m6d" Oct 01 10:43:02 crc kubenswrapper[4787]: I1001 10:43:02.129842 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e297954-d108-4f85-ab10-02634bd69bd3-utilities\") pod \"community-operators-f2m6d\" (UID: \"0e297954-d108-4f85-ab10-02634bd69bd3\") " pod="openshift-marketplace/community-operators-f2m6d" Oct 01 10:43:02 crc kubenswrapper[4787]: I1001 10:43:02.129848 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e297954-d108-4f85-ab10-02634bd69bd3-catalog-content\") pod \"community-operators-f2m6d\" (UID: \"0e297954-d108-4f85-ab10-02634bd69bd3\") " pod="openshift-marketplace/community-operators-f2m6d" Oct 01 10:43:02 crc kubenswrapper[4787]: I1001 10:43:02.157294 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmh2x\" (UniqueName: \"kubernetes.io/projected/0e297954-d108-4f85-ab10-02634bd69bd3-kube-api-access-wmh2x\") pod \"community-operators-f2m6d\" (UID: \"0e297954-d108-4f85-ab10-02634bd69bd3\") " pod="openshift-marketplace/community-operators-f2m6d" Oct 01 10:43:02 crc kubenswrapper[4787]: I1001 10:43:02.231977 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5bcd56b-9a07-4b5f-8656-0374d3b87d50-utilities\") pod \"redhat-marketplace-kftxb\" (UID: \"e5bcd56b-9a07-4b5f-8656-0374d3b87d50\") " pod="openshift-marketplace/redhat-marketplace-kftxb" Oct 01 10:43:02 crc kubenswrapper[4787]: I1001 10:43:02.232094 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5bcd56b-9a07-4b5f-8656-0374d3b87d50-catalog-content\") pod \"redhat-marketplace-kftxb\" (UID: \"e5bcd56b-9a07-4b5f-8656-0374d3b87d50\") " pod="openshift-marketplace/redhat-marketplace-kftxb" Oct 01 10:43:02 crc kubenswrapper[4787]: I1001 10:43:02.232131 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kvp8\" (UniqueName: \"kubernetes.io/projected/e5bcd56b-9a07-4b5f-8656-0374d3b87d50-kube-api-access-4kvp8\") pod \"redhat-marketplace-kftxb\" (UID: \"e5bcd56b-9a07-4b5f-8656-0374d3b87d50\") " pod="openshift-marketplace/redhat-marketplace-kftxb" Oct 01 10:43:02 crc kubenswrapper[4787]: I1001 10:43:02.233447 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f2m6d" Oct 01 10:43:02 crc kubenswrapper[4787]: I1001 10:43:02.333874 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5bcd56b-9a07-4b5f-8656-0374d3b87d50-catalog-content\") pod \"redhat-marketplace-kftxb\" (UID: \"e5bcd56b-9a07-4b5f-8656-0374d3b87d50\") " pod="openshift-marketplace/redhat-marketplace-kftxb" Oct 01 10:43:02 crc kubenswrapper[4787]: I1001 10:43:02.334246 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kvp8\" (UniqueName: \"kubernetes.io/projected/e5bcd56b-9a07-4b5f-8656-0374d3b87d50-kube-api-access-4kvp8\") pod \"redhat-marketplace-kftxb\" (UID: \"e5bcd56b-9a07-4b5f-8656-0374d3b87d50\") " pod="openshift-marketplace/redhat-marketplace-kftxb" Oct 01 10:43:02 crc kubenswrapper[4787]: I1001 10:43:02.334363 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5bcd56b-9a07-4b5f-8656-0374d3b87d50-utilities\") pod \"redhat-marketplace-kftxb\" (UID: \"e5bcd56b-9a07-4b5f-8656-0374d3b87d50\") " pod="openshift-marketplace/redhat-marketplace-kftxb" Oct 01 10:43:02 crc kubenswrapper[4787]: I1001 10:43:02.334490 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5bcd56b-9a07-4b5f-8656-0374d3b87d50-catalog-content\") pod \"redhat-marketplace-kftxb\" (UID: \"e5bcd56b-9a07-4b5f-8656-0374d3b87d50\") " pod="openshift-marketplace/redhat-marketplace-kftxb" Oct 01 10:43:02 crc kubenswrapper[4787]: I1001 10:43:02.334957 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5bcd56b-9a07-4b5f-8656-0374d3b87d50-utilities\") pod \"redhat-marketplace-kftxb\" (UID: \"e5bcd56b-9a07-4b5f-8656-0374d3b87d50\") " pod="openshift-marketplace/redhat-marketplace-kftxb" Oct 01 10:43:02 crc kubenswrapper[4787]: I1001 10:43:02.352795 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kvp8\" (UniqueName: \"kubernetes.io/projected/e5bcd56b-9a07-4b5f-8656-0374d3b87d50-kube-api-access-4kvp8\") pod \"redhat-marketplace-kftxb\" (UID: \"e5bcd56b-9a07-4b5f-8656-0374d3b87d50\") " pod="openshift-marketplace/redhat-marketplace-kftxb" Oct 01 10:43:02 crc kubenswrapper[4787]: I1001 10:43:02.433544 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kftxb" Oct 01 10:43:02 crc kubenswrapper[4787]: I1001 10:43:02.831752 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f2m6d"] Oct 01 10:43:03 crc kubenswrapper[4787]: I1001 10:43:03.003258 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kftxb"] Oct 01 10:43:03 crc kubenswrapper[4787]: W1001 10:43:03.016293 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode5bcd56b_9a07_4b5f_8656_0374d3b87d50.slice/crio-b2e7149cad61bea26176cbbae5108a0c147c778fea833b6a61c9d57d2fd357b1 WatchSource:0}: Error finding container b2e7149cad61bea26176cbbae5108a0c147c778fea833b6a61c9d57d2fd357b1: Status 404 returned error can't find the container with id b2e7149cad61bea26176cbbae5108a0c147c778fea833b6a61c9d57d2fd357b1 Oct 01 10:43:03 crc kubenswrapper[4787]: I1001 10:43:03.104234 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kftxb" event={"ID":"e5bcd56b-9a07-4b5f-8656-0374d3b87d50","Type":"ContainerStarted","Data":"b2e7149cad61bea26176cbbae5108a0c147c778fea833b6a61c9d57d2fd357b1"} Oct 01 10:43:03 crc kubenswrapper[4787]: I1001 10:43:03.106189 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f2m6d" event={"ID":"0e297954-d108-4f85-ab10-02634bd69bd3","Type":"ContainerStarted","Data":"61f2bcfc9b72321212ced83116faf2716a864fe66b4730fa9f02feb05555999b"} Oct 01 10:43:03 crc kubenswrapper[4787]: I1001 10:43:03.106291 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f2m6d" event={"ID":"0e297954-d108-4f85-ab10-02634bd69bd3","Type":"ContainerStarted","Data":"baf1e853ceec418a79691e6d2840e046deb23b60946c79f5c5a17dac15a2d227"} Oct 01 10:43:04 crc kubenswrapper[4787]: I1001 10:43:04.116234 4787 generic.go:334] "Generic (PLEG): container finished" podID="e5bcd56b-9a07-4b5f-8656-0374d3b87d50" containerID="4e14d284a027093d95c6cce3988204c6c917f17a41ed12d095c474a0c2e48b15" exitCode=0 Oct 01 10:43:04 crc kubenswrapper[4787]: I1001 10:43:04.116349 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kftxb" event={"ID":"e5bcd56b-9a07-4b5f-8656-0374d3b87d50","Type":"ContainerDied","Data":"4e14d284a027093d95c6cce3988204c6c917f17a41ed12d095c474a0c2e48b15"} Oct 01 10:43:04 crc kubenswrapper[4787]: I1001 10:43:04.118973 4787 generic.go:334] "Generic (PLEG): container finished" podID="0e297954-d108-4f85-ab10-02634bd69bd3" containerID="61f2bcfc9b72321212ced83116faf2716a864fe66b4730fa9f02feb05555999b" exitCode=0 Oct 01 10:43:04 crc kubenswrapper[4787]: I1001 10:43:04.119012 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f2m6d" event={"ID":"0e297954-d108-4f85-ab10-02634bd69bd3","Type":"ContainerDied","Data":"61f2bcfc9b72321212ced83116faf2716a864fe66b4730fa9f02feb05555999b"} Oct 01 10:43:06 crc kubenswrapper[4787]: I1001 10:43:06.142901 4787 generic.go:334] "Generic (PLEG): container finished" podID="e5bcd56b-9a07-4b5f-8656-0374d3b87d50" containerID="76862a300ec487e7198b48f4d184dcf6698faa51fee3629e64252794217bf01f" exitCode=0 Oct 01 10:43:06 crc kubenswrapper[4787]: I1001 10:43:06.142972 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kftxb" event={"ID":"e5bcd56b-9a07-4b5f-8656-0374d3b87d50","Type":"ContainerDied","Data":"76862a300ec487e7198b48f4d184dcf6698faa51fee3629e64252794217bf01f"} Oct 01 10:43:06 crc kubenswrapper[4787]: I1001 10:43:06.146807 4787 generic.go:334] "Generic (PLEG): container finished" podID="0e297954-d108-4f85-ab10-02634bd69bd3" containerID="7dbbaa9f1a526bfb84e55d009cbce26907fa95c262123d4c90d452c15b0ee049" exitCode=0 Oct 01 10:43:06 crc kubenswrapper[4787]: I1001 10:43:06.146851 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f2m6d" event={"ID":"0e297954-d108-4f85-ab10-02634bd69bd3","Type":"ContainerDied","Data":"7dbbaa9f1a526bfb84e55d009cbce26907fa95c262123d4c90d452c15b0ee049"} Oct 01 10:43:07 crc kubenswrapper[4787]: I1001 10:43:07.157861 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f2m6d" event={"ID":"0e297954-d108-4f85-ab10-02634bd69bd3","Type":"ContainerStarted","Data":"80c534999501a07130e370bd69523ba03536acc14133aadc5591f4468ff51b32"} Oct 01 10:43:07 crc kubenswrapper[4787]: I1001 10:43:07.160487 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kftxb" event={"ID":"e5bcd56b-9a07-4b5f-8656-0374d3b87d50","Type":"ContainerStarted","Data":"9622dcd3ca409608aba763fb82993a7fc3e07f1690ab5aab45d5652a15ce64b7"} Oct 01 10:43:07 crc kubenswrapper[4787]: I1001 10:43:07.184874 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-f2m6d" podStartSLOduration=3.6453018520000002 podStartE2EDuration="6.184855547s" podCreationTimestamp="2025-10-01 10:43:01 +0000 UTC" firstStartedPulling="2025-10-01 10:43:04.120143984 +0000 UTC m=+4016.235288141" lastFinishedPulling="2025-10-01 10:43:06.659697679 +0000 UTC m=+4018.774841836" observedRunningTime="2025-10-01 10:43:07.175475917 +0000 UTC m=+4019.290620084" watchObservedRunningTime="2025-10-01 10:43:07.184855547 +0000 UTC m=+4019.299999704" Oct 01 10:43:07 crc kubenswrapper[4787]: I1001 10:43:07.206606 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kftxb" podStartSLOduration=2.41011381 podStartE2EDuration="5.20658391s" podCreationTimestamp="2025-10-01 10:43:02 +0000 UTC" firstStartedPulling="2025-10-01 10:43:04.118541585 +0000 UTC m=+4016.233685742" lastFinishedPulling="2025-10-01 10:43:06.915011685 +0000 UTC m=+4019.030155842" observedRunningTime="2025-10-01 10:43:07.198693476 +0000 UTC m=+4019.313837643" watchObservedRunningTime="2025-10-01 10:43:07.20658391 +0000 UTC m=+4019.321728067" Oct 01 10:43:07 crc kubenswrapper[4787]: I1001 10:43:07.638893 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xmvqc" Oct 01 10:43:07 crc kubenswrapper[4787]: I1001 10:43:07.638939 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xmvqc" Oct 01 10:43:07 crc kubenswrapper[4787]: I1001 10:43:07.691879 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xmvqc" Oct 01 10:43:08 crc kubenswrapper[4787]: I1001 10:43:08.226314 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xmvqc" Oct 01 10:43:09 crc kubenswrapper[4787]: I1001 10:43:09.699509 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xmvqc"] Oct 01 10:43:10 crc kubenswrapper[4787]: I1001 10:43:10.187000 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xmvqc" podUID="c9f49564-58da-4338-9bff-a9c5b64b78df" containerName="registry-server" containerID="cri-o://6fdb34304eccf388d9ce6c1fcab86ad9ba44995d1fd22e017a3b3c21e6b48291" gracePeriod=2 Oct 01 10:43:11 crc kubenswrapper[4787]: I1001 10:43:11.166990 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xmvqc" Oct 01 10:43:11 crc kubenswrapper[4787]: I1001 10:43:11.198404 4787 generic.go:334] "Generic (PLEG): container finished" podID="c9f49564-58da-4338-9bff-a9c5b64b78df" containerID="6fdb34304eccf388d9ce6c1fcab86ad9ba44995d1fd22e017a3b3c21e6b48291" exitCode=0 Oct 01 10:43:11 crc kubenswrapper[4787]: I1001 10:43:11.198460 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xmvqc" event={"ID":"c9f49564-58da-4338-9bff-a9c5b64b78df","Type":"ContainerDied","Data":"6fdb34304eccf388d9ce6c1fcab86ad9ba44995d1fd22e017a3b3c21e6b48291"} Oct 01 10:43:11 crc kubenswrapper[4787]: I1001 10:43:11.198493 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xmvqc" event={"ID":"c9f49564-58da-4338-9bff-a9c5b64b78df","Type":"ContainerDied","Data":"008a8882b6eeb6192a493c576cbfcf82bdefcf7d216678c82890a5787c02304c"} Oct 01 10:43:11 crc kubenswrapper[4787]: I1001 10:43:11.198521 4787 scope.go:117] "RemoveContainer" containerID="6fdb34304eccf388d9ce6c1fcab86ad9ba44995d1fd22e017a3b3c21e6b48291" Oct 01 10:43:11 crc kubenswrapper[4787]: I1001 10:43:11.198688 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xmvqc" Oct 01 10:43:11 crc kubenswrapper[4787]: I1001 10:43:11.222890 4787 scope.go:117] "RemoveContainer" containerID="8a9a1ff5393a4b3c1f49ff34eac342bc2669c127fdddd5d6dbbccedb81175ecb" Oct 01 10:43:11 crc kubenswrapper[4787]: I1001 10:43:11.229579 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9f49564-58da-4338-9bff-a9c5b64b78df-catalog-content\") pod \"c9f49564-58da-4338-9bff-a9c5b64b78df\" (UID: \"c9f49564-58da-4338-9bff-a9c5b64b78df\") " Oct 01 10:43:11 crc kubenswrapper[4787]: I1001 10:43:11.229633 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c56lf\" (UniqueName: \"kubernetes.io/projected/c9f49564-58da-4338-9bff-a9c5b64b78df-kube-api-access-c56lf\") pod \"c9f49564-58da-4338-9bff-a9c5b64b78df\" (UID: \"c9f49564-58da-4338-9bff-a9c5b64b78df\") " Oct 01 10:43:11 crc kubenswrapper[4787]: I1001 10:43:11.229753 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9f49564-58da-4338-9bff-a9c5b64b78df-utilities\") pod \"c9f49564-58da-4338-9bff-a9c5b64b78df\" (UID: \"c9f49564-58da-4338-9bff-a9c5b64b78df\") " Oct 01 10:43:11 crc kubenswrapper[4787]: I1001 10:43:11.231895 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9f49564-58da-4338-9bff-a9c5b64b78df-utilities" (OuterVolumeSpecName: "utilities") pod "c9f49564-58da-4338-9bff-a9c5b64b78df" (UID: "c9f49564-58da-4338-9bff-a9c5b64b78df"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:43:11 crc kubenswrapper[4787]: I1001 10:43:11.244361 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9f49564-58da-4338-9bff-a9c5b64b78df-kube-api-access-c56lf" (OuterVolumeSpecName: "kube-api-access-c56lf") pod "c9f49564-58da-4338-9bff-a9c5b64b78df" (UID: "c9f49564-58da-4338-9bff-a9c5b64b78df"). InnerVolumeSpecName "kube-api-access-c56lf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:43:11 crc kubenswrapper[4787]: I1001 10:43:11.247253 4787 scope.go:117] "RemoveContainer" containerID="444d93b283bb8179cf67353e6731cddca7b5bbd1ce486a6d65bf38a73186f0f0" Oct 01 10:43:11 crc kubenswrapper[4787]: I1001 10:43:11.251150 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:43:11 crc kubenswrapper[4787]: I1001 10:43:11.251379 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:43:11 crc kubenswrapper[4787]: I1001 10:43:11.328173 4787 scope.go:117] "RemoveContainer" containerID="6fdb34304eccf388d9ce6c1fcab86ad9ba44995d1fd22e017a3b3c21e6b48291" Oct 01 10:43:11 crc kubenswrapper[4787]: E1001 10:43:11.328788 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6fdb34304eccf388d9ce6c1fcab86ad9ba44995d1fd22e017a3b3c21e6b48291\": container with ID starting with 6fdb34304eccf388d9ce6c1fcab86ad9ba44995d1fd22e017a3b3c21e6b48291 not found: ID does not exist" containerID="6fdb34304eccf388d9ce6c1fcab86ad9ba44995d1fd22e017a3b3c21e6b48291" Oct 01 10:43:11 crc kubenswrapper[4787]: I1001 10:43:11.328898 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fdb34304eccf388d9ce6c1fcab86ad9ba44995d1fd22e017a3b3c21e6b48291"} err="failed to get container status \"6fdb34304eccf388d9ce6c1fcab86ad9ba44995d1fd22e017a3b3c21e6b48291\": rpc error: code = NotFound desc = could not find container \"6fdb34304eccf388d9ce6c1fcab86ad9ba44995d1fd22e017a3b3c21e6b48291\": container with ID starting with 6fdb34304eccf388d9ce6c1fcab86ad9ba44995d1fd22e017a3b3c21e6b48291 not found: ID does not exist" Oct 01 10:43:11 crc kubenswrapper[4787]: I1001 10:43:11.328972 4787 scope.go:117] "RemoveContainer" containerID="8a9a1ff5393a4b3c1f49ff34eac342bc2669c127fdddd5d6dbbccedb81175ecb" Oct 01 10:43:11 crc kubenswrapper[4787]: E1001 10:43:11.329586 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a9a1ff5393a4b3c1f49ff34eac342bc2669c127fdddd5d6dbbccedb81175ecb\": container with ID starting with 8a9a1ff5393a4b3c1f49ff34eac342bc2669c127fdddd5d6dbbccedb81175ecb not found: ID does not exist" containerID="8a9a1ff5393a4b3c1f49ff34eac342bc2669c127fdddd5d6dbbccedb81175ecb" Oct 01 10:43:11 crc kubenswrapper[4787]: I1001 10:43:11.329694 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a9a1ff5393a4b3c1f49ff34eac342bc2669c127fdddd5d6dbbccedb81175ecb"} err="failed to get container status \"8a9a1ff5393a4b3c1f49ff34eac342bc2669c127fdddd5d6dbbccedb81175ecb\": rpc error: code = NotFound desc = could not find container \"8a9a1ff5393a4b3c1f49ff34eac342bc2669c127fdddd5d6dbbccedb81175ecb\": container with ID starting with 8a9a1ff5393a4b3c1f49ff34eac342bc2669c127fdddd5d6dbbccedb81175ecb not found: ID does not exist" Oct 01 10:43:11 crc kubenswrapper[4787]: I1001 10:43:11.329777 4787 scope.go:117] "RemoveContainer" containerID="444d93b283bb8179cf67353e6731cddca7b5bbd1ce486a6d65bf38a73186f0f0" Oct 01 10:43:11 crc kubenswrapper[4787]: E1001 10:43:11.330187 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"444d93b283bb8179cf67353e6731cddca7b5bbd1ce486a6d65bf38a73186f0f0\": container with ID starting with 444d93b283bb8179cf67353e6731cddca7b5bbd1ce486a6d65bf38a73186f0f0 not found: ID does not exist" containerID="444d93b283bb8179cf67353e6731cddca7b5bbd1ce486a6d65bf38a73186f0f0" Oct 01 10:43:11 crc kubenswrapper[4787]: I1001 10:43:11.330207 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"444d93b283bb8179cf67353e6731cddca7b5bbd1ce486a6d65bf38a73186f0f0"} err="failed to get container status \"444d93b283bb8179cf67353e6731cddca7b5bbd1ce486a6d65bf38a73186f0f0\": rpc error: code = NotFound desc = could not find container \"444d93b283bb8179cf67353e6731cddca7b5bbd1ce486a6d65bf38a73186f0f0\": container with ID starting with 444d93b283bb8179cf67353e6731cddca7b5bbd1ce486a6d65bf38a73186f0f0 not found: ID does not exist" Oct 01 10:43:11 crc kubenswrapper[4787]: I1001 10:43:11.333444 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9f49564-58da-4338-9bff-a9c5b64b78df-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 10:43:11 crc kubenswrapper[4787]: I1001 10:43:11.333483 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c56lf\" (UniqueName: \"kubernetes.io/projected/c9f49564-58da-4338-9bff-a9c5b64b78df-kube-api-access-c56lf\") on node \"crc\" DevicePath \"\"" Oct 01 10:43:12 crc kubenswrapper[4787]: I1001 10:43:12.234996 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-f2m6d" Oct 01 10:43:12 crc kubenswrapper[4787]: I1001 10:43:12.235494 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-f2m6d" Oct 01 10:43:12 crc kubenswrapper[4787]: I1001 10:43:12.283397 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-f2m6d" Oct 01 10:43:12 crc kubenswrapper[4787]: I1001 10:43:12.433824 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kftxb" Oct 01 10:43:12 crc kubenswrapper[4787]: I1001 10:43:12.433870 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kftxb" Oct 01 10:43:12 crc kubenswrapper[4787]: I1001 10:43:12.480227 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kftxb" Oct 01 10:43:12 crc kubenswrapper[4787]: I1001 10:43:12.953853 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9f49564-58da-4338-9bff-a9c5b64b78df-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c9f49564-58da-4338-9bff-a9c5b64b78df" (UID: "c9f49564-58da-4338-9bff-a9c5b64b78df"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:43:12 crc kubenswrapper[4787]: I1001 10:43:12.962538 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9f49564-58da-4338-9bff-a9c5b64b78df-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 10:43:13 crc kubenswrapper[4787]: I1001 10:43:13.047209 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xmvqc"] Oct 01 10:43:13 crc kubenswrapper[4787]: I1001 10:43:13.068874 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xmvqc"] Oct 01 10:43:13 crc kubenswrapper[4787]: I1001 10:43:13.273674 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-f2m6d" Oct 01 10:43:13 crc kubenswrapper[4787]: I1001 10:43:13.291026 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kftxb" Oct 01 10:43:14 crc kubenswrapper[4787]: I1001 10:43:14.536437 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9f49564-58da-4338-9bff-a9c5b64b78df" path="/var/lib/kubelet/pods/c9f49564-58da-4338-9bff-a9c5b64b78df/volumes" Oct 01 10:43:14 crc kubenswrapper[4787]: I1001 10:43:14.700488 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f2m6d"] Oct 01 10:43:15 crc kubenswrapper[4787]: I1001 10:43:15.304953 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kftxb"] Oct 01 10:43:15 crc kubenswrapper[4787]: I1001 10:43:15.305285 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kftxb" podUID="e5bcd56b-9a07-4b5f-8656-0374d3b87d50" containerName="registry-server" containerID="cri-o://9622dcd3ca409608aba763fb82993a7fc3e07f1690ab5aab45d5652a15ce64b7" gracePeriod=2 Oct 01 10:43:15 crc kubenswrapper[4787]: I1001 10:43:15.791159 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kftxb" Oct 01 10:43:15 crc kubenswrapper[4787]: I1001 10:43:15.832667 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5bcd56b-9a07-4b5f-8656-0374d3b87d50-utilities\") pod \"e5bcd56b-9a07-4b5f-8656-0374d3b87d50\" (UID: \"e5bcd56b-9a07-4b5f-8656-0374d3b87d50\") " Oct 01 10:43:15 crc kubenswrapper[4787]: I1001 10:43:15.832785 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5bcd56b-9a07-4b5f-8656-0374d3b87d50-catalog-content\") pod \"e5bcd56b-9a07-4b5f-8656-0374d3b87d50\" (UID: \"e5bcd56b-9a07-4b5f-8656-0374d3b87d50\") " Oct 01 10:43:15 crc kubenswrapper[4787]: I1001 10:43:15.832890 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4kvp8\" (UniqueName: \"kubernetes.io/projected/e5bcd56b-9a07-4b5f-8656-0374d3b87d50-kube-api-access-4kvp8\") pod \"e5bcd56b-9a07-4b5f-8656-0374d3b87d50\" (UID: \"e5bcd56b-9a07-4b5f-8656-0374d3b87d50\") " Oct 01 10:43:15 crc kubenswrapper[4787]: I1001 10:43:15.833523 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5bcd56b-9a07-4b5f-8656-0374d3b87d50-utilities" (OuterVolumeSpecName: "utilities") pod "e5bcd56b-9a07-4b5f-8656-0374d3b87d50" (UID: "e5bcd56b-9a07-4b5f-8656-0374d3b87d50"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:43:15 crc kubenswrapper[4787]: I1001 10:43:15.847443 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5bcd56b-9a07-4b5f-8656-0374d3b87d50-kube-api-access-4kvp8" (OuterVolumeSpecName: "kube-api-access-4kvp8") pod "e5bcd56b-9a07-4b5f-8656-0374d3b87d50" (UID: "e5bcd56b-9a07-4b5f-8656-0374d3b87d50"). InnerVolumeSpecName "kube-api-access-4kvp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:43:15 crc kubenswrapper[4787]: I1001 10:43:15.858719 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5bcd56b-9a07-4b5f-8656-0374d3b87d50-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e5bcd56b-9a07-4b5f-8656-0374d3b87d50" (UID: "e5bcd56b-9a07-4b5f-8656-0374d3b87d50"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:43:15 crc kubenswrapper[4787]: I1001 10:43:15.935757 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5bcd56b-9a07-4b5f-8656-0374d3b87d50-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 10:43:15 crc kubenswrapper[4787]: I1001 10:43:15.935789 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4kvp8\" (UniqueName: \"kubernetes.io/projected/e5bcd56b-9a07-4b5f-8656-0374d3b87d50-kube-api-access-4kvp8\") on node \"crc\" DevicePath \"\"" Oct 01 10:43:15 crc kubenswrapper[4787]: I1001 10:43:15.935799 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5bcd56b-9a07-4b5f-8656-0374d3b87d50-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 10:43:16 crc kubenswrapper[4787]: I1001 10:43:16.246021 4787 generic.go:334] "Generic (PLEG): container finished" podID="e5bcd56b-9a07-4b5f-8656-0374d3b87d50" containerID="9622dcd3ca409608aba763fb82993a7fc3e07f1690ab5aab45d5652a15ce64b7" exitCode=0 Oct 01 10:43:16 crc kubenswrapper[4787]: I1001 10:43:16.246134 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kftxb" event={"ID":"e5bcd56b-9a07-4b5f-8656-0374d3b87d50","Type":"ContainerDied","Data":"9622dcd3ca409608aba763fb82993a7fc3e07f1690ab5aab45d5652a15ce64b7"} Oct 01 10:43:16 crc kubenswrapper[4787]: I1001 10:43:16.246196 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kftxb" event={"ID":"e5bcd56b-9a07-4b5f-8656-0374d3b87d50","Type":"ContainerDied","Data":"b2e7149cad61bea26176cbbae5108a0c147c778fea833b6a61c9d57d2fd357b1"} Oct 01 10:43:16 crc kubenswrapper[4787]: I1001 10:43:16.246214 4787 scope.go:117] "RemoveContainer" containerID="9622dcd3ca409608aba763fb82993a7fc3e07f1690ab5aab45d5652a15ce64b7" Oct 01 10:43:16 crc kubenswrapper[4787]: I1001 10:43:16.246247 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kftxb" Oct 01 10:43:16 crc kubenswrapper[4787]: I1001 10:43:16.246326 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-f2m6d" podUID="0e297954-d108-4f85-ab10-02634bd69bd3" containerName="registry-server" containerID="cri-o://80c534999501a07130e370bd69523ba03536acc14133aadc5591f4468ff51b32" gracePeriod=2 Oct 01 10:43:16 crc kubenswrapper[4787]: I1001 10:43:16.272250 4787 scope.go:117] "RemoveContainer" containerID="76862a300ec487e7198b48f4d184dcf6698faa51fee3629e64252794217bf01f" Oct 01 10:43:16 crc kubenswrapper[4787]: I1001 10:43:16.310378 4787 scope.go:117] "RemoveContainer" containerID="4e14d284a027093d95c6cce3988204c6c917f17a41ed12d095c474a0c2e48b15" Oct 01 10:43:16 crc kubenswrapper[4787]: I1001 10:43:16.313370 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kftxb"] Oct 01 10:43:16 crc kubenswrapper[4787]: I1001 10:43:16.327985 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kftxb"] Oct 01 10:43:16 crc kubenswrapper[4787]: I1001 10:43:16.478034 4787 scope.go:117] "RemoveContainer" containerID="9622dcd3ca409608aba763fb82993a7fc3e07f1690ab5aab45d5652a15ce64b7" Oct 01 10:43:16 crc kubenswrapper[4787]: E1001 10:43:16.478441 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9622dcd3ca409608aba763fb82993a7fc3e07f1690ab5aab45d5652a15ce64b7\": container with ID starting with 9622dcd3ca409608aba763fb82993a7fc3e07f1690ab5aab45d5652a15ce64b7 not found: ID does not exist" containerID="9622dcd3ca409608aba763fb82993a7fc3e07f1690ab5aab45d5652a15ce64b7" Oct 01 10:43:16 crc kubenswrapper[4787]: I1001 10:43:16.478481 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9622dcd3ca409608aba763fb82993a7fc3e07f1690ab5aab45d5652a15ce64b7"} err="failed to get container status \"9622dcd3ca409608aba763fb82993a7fc3e07f1690ab5aab45d5652a15ce64b7\": rpc error: code = NotFound desc = could not find container \"9622dcd3ca409608aba763fb82993a7fc3e07f1690ab5aab45d5652a15ce64b7\": container with ID starting with 9622dcd3ca409608aba763fb82993a7fc3e07f1690ab5aab45d5652a15ce64b7 not found: ID does not exist" Oct 01 10:43:16 crc kubenswrapper[4787]: I1001 10:43:16.478503 4787 scope.go:117] "RemoveContainer" containerID="76862a300ec487e7198b48f4d184dcf6698faa51fee3629e64252794217bf01f" Oct 01 10:43:16 crc kubenswrapper[4787]: E1001 10:43:16.478812 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76862a300ec487e7198b48f4d184dcf6698faa51fee3629e64252794217bf01f\": container with ID starting with 76862a300ec487e7198b48f4d184dcf6698faa51fee3629e64252794217bf01f not found: ID does not exist" containerID="76862a300ec487e7198b48f4d184dcf6698faa51fee3629e64252794217bf01f" Oct 01 10:43:16 crc kubenswrapper[4787]: I1001 10:43:16.478866 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76862a300ec487e7198b48f4d184dcf6698faa51fee3629e64252794217bf01f"} err="failed to get container status \"76862a300ec487e7198b48f4d184dcf6698faa51fee3629e64252794217bf01f\": rpc error: code = NotFound desc = could not find container \"76862a300ec487e7198b48f4d184dcf6698faa51fee3629e64252794217bf01f\": container with ID starting with 76862a300ec487e7198b48f4d184dcf6698faa51fee3629e64252794217bf01f not found: ID does not exist" Oct 01 10:43:16 crc kubenswrapper[4787]: I1001 10:43:16.478900 4787 scope.go:117] "RemoveContainer" containerID="4e14d284a027093d95c6cce3988204c6c917f17a41ed12d095c474a0c2e48b15" Oct 01 10:43:16 crc kubenswrapper[4787]: E1001 10:43:16.479327 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e14d284a027093d95c6cce3988204c6c917f17a41ed12d095c474a0c2e48b15\": container with ID starting with 4e14d284a027093d95c6cce3988204c6c917f17a41ed12d095c474a0c2e48b15 not found: ID does not exist" containerID="4e14d284a027093d95c6cce3988204c6c917f17a41ed12d095c474a0c2e48b15" Oct 01 10:43:16 crc kubenswrapper[4787]: I1001 10:43:16.479371 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e14d284a027093d95c6cce3988204c6c917f17a41ed12d095c474a0c2e48b15"} err="failed to get container status \"4e14d284a027093d95c6cce3988204c6c917f17a41ed12d095c474a0c2e48b15\": rpc error: code = NotFound desc = could not find container \"4e14d284a027093d95c6cce3988204c6c917f17a41ed12d095c474a0c2e48b15\": container with ID starting with 4e14d284a027093d95c6cce3988204c6c917f17a41ed12d095c474a0c2e48b15 not found: ID does not exist" Oct 01 10:43:16 crc kubenswrapper[4787]: I1001 10:43:16.545762 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5bcd56b-9a07-4b5f-8656-0374d3b87d50" path="/var/lib/kubelet/pods/e5bcd56b-9a07-4b5f-8656-0374d3b87d50/volumes" Oct 01 10:43:16 crc kubenswrapper[4787]: I1001 10:43:16.701700 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f2m6d" Oct 01 10:43:16 crc kubenswrapper[4787]: I1001 10:43:16.751886 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e297954-d108-4f85-ab10-02634bd69bd3-catalog-content\") pod \"0e297954-d108-4f85-ab10-02634bd69bd3\" (UID: \"0e297954-d108-4f85-ab10-02634bd69bd3\") " Oct 01 10:43:16 crc kubenswrapper[4787]: I1001 10:43:16.752029 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e297954-d108-4f85-ab10-02634bd69bd3-utilities\") pod \"0e297954-d108-4f85-ab10-02634bd69bd3\" (UID: \"0e297954-d108-4f85-ab10-02634bd69bd3\") " Oct 01 10:43:16 crc kubenswrapper[4787]: I1001 10:43:16.752070 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wmh2x\" (UniqueName: \"kubernetes.io/projected/0e297954-d108-4f85-ab10-02634bd69bd3-kube-api-access-wmh2x\") pod \"0e297954-d108-4f85-ab10-02634bd69bd3\" (UID: \"0e297954-d108-4f85-ab10-02634bd69bd3\") " Oct 01 10:43:16 crc kubenswrapper[4787]: I1001 10:43:16.760351 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e297954-d108-4f85-ab10-02634bd69bd3-kube-api-access-wmh2x" (OuterVolumeSpecName: "kube-api-access-wmh2x") pod "0e297954-d108-4f85-ab10-02634bd69bd3" (UID: "0e297954-d108-4f85-ab10-02634bd69bd3"). InnerVolumeSpecName "kube-api-access-wmh2x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:43:16 crc kubenswrapper[4787]: I1001 10:43:16.760570 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e297954-d108-4f85-ab10-02634bd69bd3-utilities" (OuterVolumeSpecName: "utilities") pod "0e297954-d108-4f85-ab10-02634bd69bd3" (UID: "0e297954-d108-4f85-ab10-02634bd69bd3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:43:16 crc kubenswrapper[4787]: I1001 10:43:16.805593 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e297954-d108-4f85-ab10-02634bd69bd3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0e297954-d108-4f85-ab10-02634bd69bd3" (UID: "0e297954-d108-4f85-ab10-02634bd69bd3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:43:16 crc kubenswrapper[4787]: I1001 10:43:16.869349 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e297954-d108-4f85-ab10-02634bd69bd3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 01 10:43:16 crc kubenswrapper[4787]: I1001 10:43:16.869852 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e297954-d108-4f85-ab10-02634bd69bd3-utilities\") on node \"crc\" DevicePath \"\"" Oct 01 10:43:16 crc kubenswrapper[4787]: I1001 10:43:16.869911 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wmh2x\" (UniqueName: \"kubernetes.io/projected/0e297954-d108-4f85-ab10-02634bd69bd3-kube-api-access-wmh2x\") on node \"crc\" DevicePath \"\"" Oct 01 10:43:17 crc kubenswrapper[4787]: I1001 10:43:17.260491 4787 generic.go:334] "Generic (PLEG): container finished" podID="0e297954-d108-4f85-ab10-02634bd69bd3" containerID="80c534999501a07130e370bd69523ba03536acc14133aadc5591f4468ff51b32" exitCode=0 Oct 01 10:43:17 crc kubenswrapper[4787]: I1001 10:43:17.260563 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f2m6d" Oct 01 10:43:17 crc kubenswrapper[4787]: I1001 10:43:17.260595 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f2m6d" event={"ID":"0e297954-d108-4f85-ab10-02634bd69bd3","Type":"ContainerDied","Data":"80c534999501a07130e370bd69523ba03536acc14133aadc5591f4468ff51b32"} Oct 01 10:43:17 crc kubenswrapper[4787]: I1001 10:43:17.260669 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f2m6d" event={"ID":"0e297954-d108-4f85-ab10-02634bd69bd3","Type":"ContainerDied","Data":"baf1e853ceec418a79691e6d2840e046deb23b60946c79f5c5a17dac15a2d227"} Oct 01 10:43:17 crc kubenswrapper[4787]: I1001 10:43:17.260688 4787 scope.go:117] "RemoveContainer" containerID="80c534999501a07130e370bd69523ba03536acc14133aadc5591f4468ff51b32" Oct 01 10:43:17 crc kubenswrapper[4787]: I1001 10:43:17.300903 4787 scope.go:117] "RemoveContainer" containerID="7dbbaa9f1a526bfb84e55d009cbce26907fa95c262123d4c90d452c15b0ee049" Oct 01 10:43:17 crc kubenswrapper[4787]: I1001 10:43:17.306274 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f2m6d"] Oct 01 10:43:17 crc kubenswrapper[4787]: I1001 10:43:17.322846 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-f2m6d"] Oct 01 10:43:17 crc kubenswrapper[4787]: I1001 10:43:17.333327 4787 scope.go:117] "RemoveContainer" containerID="61f2bcfc9b72321212ced83116faf2716a864fe66b4730fa9f02feb05555999b" Oct 01 10:43:17 crc kubenswrapper[4787]: I1001 10:43:17.355499 4787 scope.go:117] "RemoveContainer" containerID="80c534999501a07130e370bd69523ba03536acc14133aadc5591f4468ff51b32" Oct 01 10:43:17 crc kubenswrapper[4787]: E1001 10:43:17.356195 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80c534999501a07130e370bd69523ba03536acc14133aadc5591f4468ff51b32\": container with ID starting with 80c534999501a07130e370bd69523ba03536acc14133aadc5591f4468ff51b32 not found: ID does not exist" containerID="80c534999501a07130e370bd69523ba03536acc14133aadc5591f4468ff51b32" Oct 01 10:43:17 crc kubenswrapper[4787]: I1001 10:43:17.356234 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80c534999501a07130e370bd69523ba03536acc14133aadc5591f4468ff51b32"} err="failed to get container status \"80c534999501a07130e370bd69523ba03536acc14133aadc5591f4468ff51b32\": rpc error: code = NotFound desc = could not find container \"80c534999501a07130e370bd69523ba03536acc14133aadc5591f4468ff51b32\": container with ID starting with 80c534999501a07130e370bd69523ba03536acc14133aadc5591f4468ff51b32 not found: ID does not exist" Oct 01 10:43:17 crc kubenswrapper[4787]: I1001 10:43:17.356260 4787 scope.go:117] "RemoveContainer" containerID="7dbbaa9f1a526bfb84e55d009cbce26907fa95c262123d4c90d452c15b0ee049" Oct 01 10:43:17 crc kubenswrapper[4787]: E1001 10:43:17.356690 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7dbbaa9f1a526bfb84e55d009cbce26907fa95c262123d4c90d452c15b0ee049\": container with ID starting with 7dbbaa9f1a526bfb84e55d009cbce26907fa95c262123d4c90d452c15b0ee049 not found: ID does not exist" containerID="7dbbaa9f1a526bfb84e55d009cbce26907fa95c262123d4c90d452c15b0ee049" Oct 01 10:43:17 crc kubenswrapper[4787]: I1001 10:43:17.356752 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7dbbaa9f1a526bfb84e55d009cbce26907fa95c262123d4c90d452c15b0ee049"} err="failed to get container status \"7dbbaa9f1a526bfb84e55d009cbce26907fa95c262123d4c90d452c15b0ee049\": rpc error: code = NotFound desc = could not find container \"7dbbaa9f1a526bfb84e55d009cbce26907fa95c262123d4c90d452c15b0ee049\": container with ID starting with 7dbbaa9f1a526bfb84e55d009cbce26907fa95c262123d4c90d452c15b0ee049 not found: ID does not exist" Oct 01 10:43:17 crc kubenswrapper[4787]: I1001 10:43:17.356780 4787 scope.go:117] "RemoveContainer" containerID="61f2bcfc9b72321212ced83116faf2716a864fe66b4730fa9f02feb05555999b" Oct 01 10:43:17 crc kubenswrapper[4787]: E1001 10:43:17.357295 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61f2bcfc9b72321212ced83116faf2716a864fe66b4730fa9f02feb05555999b\": container with ID starting with 61f2bcfc9b72321212ced83116faf2716a864fe66b4730fa9f02feb05555999b not found: ID does not exist" containerID="61f2bcfc9b72321212ced83116faf2716a864fe66b4730fa9f02feb05555999b" Oct 01 10:43:17 crc kubenswrapper[4787]: I1001 10:43:17.357326 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61f2bcfc9b72321212ced83116faf2716a864fe66b4730fa9f02feb05555999b"} err="failed to get container status \"61f2bcfc9b72321212ced83116faf2716a864fe66b4730fa9f02feb05555999b\": rpc error: code = NotFound desc = could not find container \"61f2bcfc9b72321212ced83116faf2716a864fe66b4730fa9f02feb05555999b\": container with ID starting with 61f2bcfc9b72321212ced83116faf2716a864fe66b4730fa9f02feb05555999b not found: ID does not exist" Oct 01 10:43:18 crc kubenswrapper[4787]: I1001 10:43:18.536606 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e297954-d108-4f85-ab10-02634bd69bd3" path="/var/lib/kubelet/pods/0e297954-d108-4f85-ab10-02634bd69bd3/volumes" Oct 01 10:43:41 crc kubenswrapper[4787]: I1001 10:43:41.250979 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:43:41 crc kubenswrapper[4787]: I1001 10:43:41.251730 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:43:41 crc kubenswrapper[4787]: I1001 10:43:41.251808 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" Oct 01 10:43:41 crc kubenswrapper[4787]: I1001 10:43:41.253165 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6e2358915272c02486736120944ef044ca65ab4dad3e49439ebe14ee8e2531f0"} pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 01 10:43:41 crc kubenswrapper[4787]: I1001 10:43:41.253302 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" containerID="cri-o://6e2358915272c02486736120944ef044ca65ab4dad3e49439ebe14ee8e2531f0" gracePeriod=600 Oct 01 10:43:41 crc kubenswrapper[4787]: I1001 10:43:41.499037 4787 generic.go:334] "Generic (PLEG): container finished" podID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerID="6e2358915272c02486736120944ef044ca65ab4dad3e49439ebe14ee8e2531f0" exitCode=0 Oct 01 10:43:41 crc kubenswrapper[4787]: I1001 10:43:41.499087 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" event={"ID":"fa51b883-5f3e-4141-9d4c-37704aac7718","Type":"ContainerDied","Data":"6e2358915272c02486736120944ef044ca65ab4dad3e49439ebe14ee8e2531f0"} Oct 01 10:43:41 crc kubenswrapper[4787]: I1001 10:43:41.499149 4787 scope.go:117] "RemoveContainer" containerID="7cb2890ef30bbede4935afca82be38385def845a6f5b1c0bd5934ac785342b56" Oct 01 10:43:42 crc kubenswrapper[4787]: I1001 10:43:42.511517 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" event={"ID":"fa51b883-5f3e-4141-9d4c-37704aac7718","Type":"ContainerStarted","Data":"92cf66caed5be2b1780cacf64612e100b202415fd1c27f800fdb71c5a672822c"} Oct 01 10:44:23 crc kubenswrapper[4787]: I1001 10:44:23.945622 4787 generic.go:334] "Generic (PLEG): container finished" podID="32eecbc7-b51a-465f-84ce-5352da7cf52a" containerID="7922e468653e12e1e904dbd354bc84e8bae31b0489b42328fb07f7d8b6f42647" exitCode=0 Oct 01 10:44:23 crc kubenswrapper[4787]: I1001 10:44:23.945688 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x2bwg/must-gather-ht4gk" event={"ID":"32eecbc7-b51a-465f-84ce-5352da7cf52a","Type":"ContainerDied","Data":"7922e468653e12e1e904dbd354bc84e8bae31b0489b42328fb07f7d8b6f42647"} Oct 01 10:44:23 crc kubenswrapper[4787]: I1001 10:44:23.947968 4787 scope.go:117] "RemoveContainer" containerID="7922e468653e12e1e904dbd354bc84e8bae31b0489b42328fb07f7d8b6f42647" Oct 01 10:44:24 crc kubenswrapper[4787]: I1001 10:44:24.796614 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-x2bwg_must-gather-ht4gk_32eecbc7-b51a-465f-84ce-5352da7cf52a/gather/0.log" Oct 01 10:44:35 crc kubenswrapper[4787]: I1001 10:44:35.964975 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-x2bwg/must-gather-ht4gk"] Oct 01 10:44:35 crc kubenswrapper[4787]: I1001 10:44:35.966101 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-x2bwg/must-gather-ht4gk" podUID="32eecbc7-b51a-465f-84ce-5352da7cf52a" containerName="copy" containerID="cri-o://d5bdcdb73c47397f098127324e8996a20cb0e14f524c114cdb1b4dc0ebc11b08" gracePeriod=2 Oct 01 10:44:35 crc kubenswrapper[4787]: I1001 10:44:35.977753 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-x2bwg/must-gather-ht4gk"] Oct 01 10:44:36 crc kubenswrapper[4787]: I1001 10:44:36.529521 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-x2bwg_must-gather-ht4gk_32eecbc7-b51a-465f-84ce-5352da7cf52a/copy/0.log" Oct 01 10:44:36 crc kubenswrapper[4787]: I1001 10:44:36.530394 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x2bwg/must-gather-ht4gk" Oct 01 10:44:36 crc kubenswrapper[4787]: I1001 10:44:36.551832 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/32eecbc7-b51a-465f-84ce-5352da7cf52a-must-gather-output\") pod \"32eecbc7-b51a-465f-84ce-5352da7cf52a\" (UID: \"32eecbc7-b51a-465f-84ce-5352da7cf52a\") " Oct 01 10:44:36 crc kubenswrapper[4787]: I1001 10:44:36.557554 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ft755\" (UniqueName: \"kubernetes.io/projected/32eecbc7-b51a-465f-84ce-5352da7cf52a-kube-api-access-ft755\") pod \"32eecbc7-b51a-465f-84ce-5352da7cf52a\" (UID: \"32eecbc7-b51a-465f-84ce-5352da7cf52a\") " Oct 01 10:44:36 crc kubenswrapper[4787]: I1001 10:44:36.569560 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32eecbc7-b51a-465f-84ce-5352da7cf52a-kube-api-access-ft755" (OuterVolumeSpecName: "kube-api-access-ft755") pod "32eecbc7-b51a-465f-84ce-5352da7cf52a" (UID: "32eecbc7-b51a-465f-84ce-5352da7cf52a"). InnerVolumeSpecName "kube-api-access-ft755". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:44:36 crc kubenswrapper[4787]: I1001 10:44:36.660667 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ft755\" (UniqueName: \"kubernetes.io/projected/32eecbc7-b51a-465f-84ce-5352da7cf52a-kube-api-access-ft755\") on node \"crc\" DevicePath \"\"" Oct 01 10:44:36 crc kubenswrapper[4787]: I1001 10:44:36.709249 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32eecbc7-b51a-465f-84ce-5352da7cf52a-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "32eecbc7-b51a-465f-84ce-5352da7cf52a" (UID: "32eecbc7-b51a-465f-84ce-5352da7cf52a"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 01 10:44:36 crc kubenswrapper[4787]: I1001 10:44:36.762575 4787 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/32eecbc7-b51a-465f-84ce-5352da7cf52a-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 01 10:44:37 crc kubenswrapper[4787]: I1001 10:44:37.090160 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-x2bwg_must-gather-ht4gk_32eecbc7-b51a-465f-84ce-5352da7cf52a/copy/0.log" Oct 01 10:44:37 crc kubenswrapper[4787]: I1001 10:44:37.091258 4787 generic.go:334] "Generic (PLEG): container finished" podID="32eecbc7-b51a-465f-84ce-5352da7cf52a" containerID="d5bdcdb73c47397f098127324e8996a20cb0e14f524c114cdb1b4dc0ebc11b08" exitCode=143 Oct 01 10:44:37 crc kubenswrapper[4787]: I1001 10:44:37.091381 4787 scope.go:117] "RemoveContainer" containerID="d5bdcdb73c47397f098127324e8996a20cb0e14f524c114cdb1b4dc0ebc11b08" Oct 01 10:44:37 crc kubenswrapper[4787]: I1001 10:44:37.091426 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x2bwg/must-gather-ht4gk" Oct 01 10:44:37 crc kubenswrapper[4787]: I1001 10:44:37.121733 4787 scope.go:117] "RemoveContainer" containerID="7922e468653e12e1e904dbd354bc84e8bae31b0489b42328fb07f7d8b6f42647" Oct 01 10:44:37 crc kubenswrapper[4787]: I1001 10:44:37.303770 4787 scope.go:117] "RemoveContainer" containerID="d5bdcdb73c47397f098127324e8996a20cb0e14f524c114cdb1b4dc0ebc11b08" Oct 01 10:44:37 crc kubenswrapper[4787]: E1001 10:44:37.304613 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5bdcdb73c47397f098127324e8996a20cb0e14f524c114cdb1b4dc0ebc11b08\": container with ID starting with d5bdcdb73c47397f098127324e8996a20cb0e14f524c114cdb1b4dc0ebc11b08 not found: ID does not exist" containerID="d5bdcdb73c47397f098127324e8996a20cb0e14f524c114cdb1b4dc0ebc11b08" Oct 01 10:44:37 crc kubenswrapper[4787]: I1001 10:44:37.304741 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5bdcdb73c47397f098127324e8996a20cb0e14f524c114cdb1b4dc0ebc11b08"} err="failed to get container status \"d5bdcdb73c47397f098127324e8996a20cb0e14f524c114cdb1b4dc0ebc11b08\": rpc error: code = NotFound desc = could not find container \"d5bdcdb73c47397f098127324e8996a20cb0e14f524c114cdb1b4dc0ebc11b08\": container with ID starting with d5bdcdb73c47397f098127324e8996a20cb0e14f524c114cdb1b4dc0ebc11b08 not found: ID does not exist" Oct 01 10:44:37 crc kubenswrapper[4787]: I1001 10:44:37.304864 4787 scope.go:117] "RemoveContainer" containerID="7922e468653e12e1e904dbd354bc84e8bae31b0489b42328fb07f7d8b6f42647" Oct 01 10:44:37 crc kubenswrapper[4787]: E1001 10:44:37.305439 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7922e468653e12e1e904dbd354bc84e8bae31b0489b42328fb07f7d8b6f42647\": container with ID starting with 7922e468653e12e1e904dbd354bc84e8bae31b0489b42328fb07f7d8b6f42647 not found: ID does not exist" containerID="7922e468653e12e1e904dbd354bc84e8bae31b0489b42328fb07f7d8b6f42647" Oct 01 10:44:37 crc kubenswrapper[4787]: I1001 10:44:37.305541 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7922e468653e12e1e904dbd354bc84e8bae31b0489b42328fb07f7d8b6f42647"} err="failed to get container status \"7922e468653e12e1e904dbd354bc84e8bae31b0489b42328fb07f7d8b6f42647\": rpc error: code = NotFound desc = could not find container \"7922e468653e12e1e904dbd354bc84e8bae31b0489b42328fb07f7d8b6f42647\": container with ID starting with 7922e468653e12e1e904dbd354bc84e8bae31b0489b42328fb07f7d8b6f42647 not found: ID does not exist" Oct 01 10:44:38 crc kubenswrapper[4787]: I1001 10:44:38.543716 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32eecbc7-b51a-465f-84ce-5352da7cf52a" path="/var/lib/kubelet/pods/32eecbc7-b51a-465f-84ce-5352da7cf52a/volumes" Oct 01 10:45:00 crc kubenswrapper[4787]: I1001 10:45:00.160807 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321925-nm6wd"] Oct 01 10:45:00 crc kubenswrapper[4787]: E1001 10:45:00.161972 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e297954-d108-4f85-ab10-02634bd69bd3" containerName="extract-content" Oct 01 10:45:00 crc kubenswrapper[4787]: I1001 10:45:00.161986 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e297954-d108-4f85-ab10-02634bd69bd3" containerName="extract-content" Oct 01 10:45:00 crc kubenswrapper[4787]: E1001 10:45:00.162037 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9f49564-58da-4338-9bff-a9c5b64b78df" containerName="registry-server" Oct 01 10:45:00 crc kubenswrapper[4787]: I1001 10:45:00.162044 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9f49564-58da-4338-9bff-a9c5b64b78df" containerName="registry-server" Oct 01 10:45:00 crc kubenswrapper[4787]: E1001 10:45:00.162061 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5bcd56b-9a07-4b5f-8656-0374d3b87d50" containerName="extract-content" Oct 01 10:45:00 crc kubenswrapper[4787]: I1001 10:45:00.162068 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5bcd56b-9a07-4b5f-8656-0374d3b87d50" containerName="extract-content" Oct 01 10:45:00 crc kubenswrapper[4787]: E1001 10:45:00.162104 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e297954-d108-4f85-ab10-02634bd69bd3" containerName="registry-server" Oct 01 10:45:00 crc kubenswrapper[4787]: I1001 10:45:00.162111 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e297954-d108-4f85-ab10-02634bd69bd3" containerName="registry-server" Oct 01 10:45:00 crc kubenswrapper[4787]: E1001 10:45:00.162126 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5bcd56b-9a07-4b5f-8656-0374d3b87d50" containerName="registry-server" Oct 01 10:45:00 crc kubenswrapper[4787]: I1001 10:45:00.162131 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5bcd56b-9a07-4b5f-8656-0374d3b87d50" containerName="registry-server" Oct 01 10:45:00 crc kubenswrapper[4787]: E1001 10:45:00.162142 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e297954-d108-4f85-ab10-02634bd69bd3" containerName="extract-utilities" Oct 01 10:45:00 crc kubenswrapper[4787]: I1001 10:45:00.162148 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e297954-d108-4f85-ab10-02634bd69bd3" containerName="extract-utilities" Oct 01 10:45:00 crc kubenswrapper[4787]: E1001 10:45:00.162157 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9f49564-58da-4338-9bff-a9c5b64b78df" containerName="extract-utilities" Oct 01 10:45:00 crc kubenswrapper[4787]: I1001 10:45:00.162162 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9f49564-58da-4338-9bff-a9c5b64b78df" containerName="extract-utilities" Oct 01 10:45:00 crc kubenswrapper[4787]: E1001 10:45:00.162168 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9f49564-58da-4338-9bff-a9c5b64b78df" containerName="extract-content" Oct 01 10:45:00 crc kubenswrapper[4787]: I1001 10:45:00.162175 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9f49564-58da-4338-9bff-a9c5b64b78df" containerName="extract-content" Oct 01 10:45:00 crc kubenswrapper[4787]: E1001 10:45:00.162184 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5bcd56b-9a07-4b5f-8656-0374d3b87d50" containerName="extract-utilities" Oct 01 10:45:00 crc kubenswrapper[4787]: I1001 10:45:00.162191 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5bcd56b-9a07-4b5f-8656-0374d3b87d50" containerName="extract-utilities" Oct 01 10:45:00 crc kubenswrapper[4787]: E1001 10:45:00.162205 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32eecbc7-b51a-465f-84ce-5352da7cf52a" containerName="copy" Oct 01 10:45:00 crc kubenswrapper[4787]: I1001 10:45:00.162210 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="32eecbc7-b51a-465f-84ce-5352da7cf52a" containerName="copy" Oct 01 10:45:00 crc kubenswrapper[4787]: E1001 10:45:00.162224 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32eecbc7-b51a-465f-84ce-5352da7cf52a" containerName="gather" Oct 01 10:45:00 crc kubenswrapper[4787]: I1001 10:45:00.162230 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="32eecbc7-b51a-465f-84ce-5352da7cf52a" containerName="gather" Oct 01 10:45:00 crc kubenswrapper[4787]: I1001 10:45:00.162426 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9f49564-58da-4338-9bff-a9c5b64b78df" containerName="registry-server" Oct 01 10:45:00 crc kubenswrapper[4787]: I1001 10:45:00.162441 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5bcd56b-9a07-4b5f-8656-0374d3b87d50" containerName="registry-server" Oct 01 10:45:00 crc kubenswrapper[4787]: I1001 10:45:00.162461 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e297954-d108-4f85-ab10-02634bd69bd3" containerName="registry-server" Oct 01 10:45:00 crc kubenswrapper[4787]: I1001 10:45:00.162484 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="32eecbc7-b51a-465f-84ce-5352da7cf52a" containerName="copy" Oct 01 10:45:00 crc kubenswrapper[4787]: I1001 10:45:00.162494 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="32eecbc7-b51a-465f-84ce-5352da7cf52a" containerName="gather" Oct 01 10:45:00 crc kubenswrapper[4787]: I1001 10:45:00.163312 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321925-nm6wd" Oct 01 10:45:00 crc kubenswrapper[4787]: I1001 10:45:00.166955 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 01 10:45:00 crc kubenswrapper[4787]: I1001 10:45:00.166978 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 01 10:45:00 crc kubenswrapper[4787]: I1001 10:45:00.173688 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321925-nm6wd"] Oct 01 10:45:00 crc kubenswrapper[4787]: I1001 10:45:00.335272 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5feed863-86f3-46eb-811c-f57daec37da6-secret-volume\") pod \"collect-profiles-29321925-nm6wd\" (UID: \"5feed863-86f3-46eb-811c-f57daec37da6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321925-nm6wd" Oct 01 10:45:00 crc kubenswrapper[4787]: I1001 10:45:00.335710 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5feed863-86f3-46eb-811c-f57daec37da6-config-volume\") pod \"collect-profiles-29321925-nm6wd\" (UID: \"5feed863-86f3-46eb-811c-f57daec37da6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321925-nm6wd" Oct 01 10:45:00 crc kubenswrapper[4787]: I1001 10:45:00.335791 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75grv\" (UniqueName: \"kubernetes.io/projected/5feed863-86f3-46eb-811c-f57daec37da6-kube-api-access-75grv\") pod \"collect-profiles-29321925-nm6wd\" (UID: \"5feed863-86f3-46eb-811c-f57daec37da6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321925-nm6wd" Oct 01 10:45:00 crc kubenswrapper[4787]: I1001 10:45:00.437418 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5feed863-86f3-46eb-811c-f57daec37da6-config-volume\") pod \"collect-profiles-29321925-nm6wd\" (UID: \"5feed863-86f3-46eb-811c-f57daec37da6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321925-nm6wd" Oct 01 10:45:00 crc kubenswrapper[4787]: I1001 10:45:00.437587 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75grv\" (UniqueName: \"kubernetes.io/projected/5feed863-86f3-46eb-811c-f57daec37da6-kube-api-access-75grv\") pod \"collect-profiles-29321925-nm6wd\" (UID: \"5feed863-86f3-46eb-811c-f57daec37da6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321925-nm6wd" Oct 01 10:45:00 crc kubenswrapper[4787]: I1001 10:45:00.437678 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5feed863-86f3-46eb-811c-f57daec37da6-secret-volume\") pod \"collect-profiles-29321925-nm6wd\" (UID: \"5feed863-86f3-46eb-811c-f57daec37da6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321925-nm6wd" Oct 01 10:45:00 crc kubenswrapper[4787]: I1001 10:45:00.439937 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5feed863-86f3-46eb-811c-f57daec37da6-config-volume\") pod \"collect-profiles-29321925-nm6wd\" (UID: \"5feed863-86f3-46eb-811c-f57daec37da6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321925-nm6wd" Oct 01 10:45:00 crc kubenswrapper[4787]: I1001 10:45:00.455507 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5feed863-86f3-46eb-811c-f57daec37da6-secret-volume\") pod \"collect-profiles-29321925-nm6wd\" (UID: \"5feed863-86f3-46eb-811c-f57daec37da6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321925-nm6wd" Oct 01 10:45:00 crc kubenswrapper[4787]: I1001 10:45:00.456198 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75grv\" (UniqueName: \"kubernetes.io/projected/5feed863-86f3-46eb-811c-f57daec37da6-kube-api-access-75grv\") pod \"collect-profiles-29321925-nm6wd\" (UID: \"5feed863-86f3-46eb-811c-f57daec37da6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29321925-nm6wd" Oct 01 10:45:00 crc kubenswrapper[4787]: I1001 10:45:00.505281 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321925-nm6wd" Oct 01 10:45:00 crc kubenswrapper[4787]: I1001 10:45:00.983129 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321925-nm6wd"] Oct 01 10:45:01 crc kubenswrapper[4787]: I1001 10:45:01.372458 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321925-nm6wd" event={"ID":"5feed863-86f3-46eb-811c-f57daec37da6","Type":"ContainerStarted","Data":"e20694b45029fb7e3d3e61b6abf8961d0e5f43531d5e13896c26e86b169cbf1b"} Oct 01 10:45:01 crc kubenswrapper[4787]: I1001 10:45:01.372823 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321925-nm6wd" event={"ID":"5feed863-86f3-46eb-811c-f57daec37da6","Type":"ContainerStarted","Data":"f28894ca2b19f1ed8167b1038b9863274e9c512b4957bb3fa094b3384519fa1c"} Oct 01 10:45:01 crc kubenswrapper[4787]: I1001 10:45:01.396069 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29321925-nm6wd" podStartSLOduration=1.3960487320000001 podStartE2EDuration="1.396048732s" podCreationTimestamp="2025-10-01 10:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-01 10:45:01.383737081 +0000 UTC m=+4133.498881258" watchObservedRunningTime="2025-10-01 10:45:01.396048732 +0000 UTC m=+4133.511192899" Oct 01 10:45:02 crc kubenswrapper[4787]: I1001 10:45:02.392260 4787 generic.go:334] "Generic (PLEG): container finished" podID="5feed863-86f3-46eb-811c-f57daec37da6" containerID="e20694b45029fb7e3d3e61b6abf8961d0e5f43531d5e13896c26e86b169cbf1b" exitCode=0 Oct 01 10:45:02 crc kubenswrapper[4787]: I1001 10:45:02.392754 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321925-nm6wd" event={"ID":"5feed863-86f3-46eb-811c-f57daec37da6","Type":"ContainerDied","Data":"e20694b45029fb7e3d3e61b6abf8961d0e5f43531d5e13896c26e86b169cbf1b"} Oct 01 10:45:03 crc kubenswrapper[4787]: I1001 10:45:03.681614 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321925-nm6wd" Oct 01 10:45:03 crc kubenswrapper[4787]: I1001 10:45:03.807867 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5feed863-86f3-46eb-811c-f57daec37da6-secret-volume\") pod \"5feed863-86f3-46eb-811c-f57daec37da6\" (UID: \"5feed863-86f3-46eb-811c-f57daec37da6\") " Oct 01 10:45:03 crc kubenswrapper[4787]: I1001 10:45:03.808005 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75grv\" (UniqueName: \"kubernetes.io/projected/5feed863-86f3-46eb-811c-f57daec37da6-kube-api-access-75grv\") pod \"5feed863-86f3-46eb-811c-f57daec37da6\" (UID: \"5feed863-86f3-46eb-811c-f57daec37da6\") " Oct 01 10:45:03 crc kubenswrapper[4787]: I1001 10:45:03.808028 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5feed863-86f3-46eb-811c-f57daec37da6-config-volume\") pod \"5feed863-86f3-46eb-811c-f57daec37da6\" (UID: \"5feed863-86f3-46eb-811c-f57daec37da6\") " Oct 01 10:45:03 crc kubenswrapper[4787]: I1001 10:45:03.810288 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5feed863-86f3-46eb-811c-f57daec37da6-config-volume" (OuterVolumeSpecName: "config-volume") pod "5feed863-86f3-46eb-811c-f57daec37da6" (UID: "5feed863-86f3-46eb-811c-f57daec37da6"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 01 10:45:03 crc kubenswrapper[4787]: I1001 10:45:03.815486 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5feed863-86f3-46eb-811c-f57daec37da6-kube-api-access-75grv" (OuterVolumeSpecName: "kube-api-access-75grv") pod "5feed863-86f3-46eb-811c-f57daec37da6" (UID: "5feed863-86f3-46eb-811c-f57daec37da6"). InnerVolumeSpecName "kube-api-access-75grv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 01 10:45:03 crc kubenswrapper[4787]: I1001 10:45:03.815909 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5feed863-86f3-46eb-811c-f57daec37da6-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5feed863-86f3-46eb-811c-f57daec37da6" (UID: "5feed863-86f3-46eb-811c-f57daec37da6"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 01 10:45:03 crc kubenswrapper[4787]: I1001 10:45:03.911190 4787 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5feed863-86f3-46eb-811c-f57daec37da6-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 01 10:45:03 crc kubenswrapper[4787]: I1001 10:45:03.911249 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75grv\" (UniqueName: \"kubernetes.io/projected/5feed863-86f3-46eb-811c-f57daec37da6-kube-api-access-75grv\") on node \"crc\" DevicePath \"\"" Oct 01 10:45:03 crc kubenswrapper[4787]: I1001 10:45:03.911287 4787 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5feed863-86f3-46eb-811c-f57daec37da6-config-volume\") on node \"crc\" DevicePath \"\"" Oct 01 10:45:04 crc kubenswrapper[4787]: I1001 10:45:04.430954 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29321925-nm6wd" event={"ID":"5feed863-86f3-46eb-811c-f57daec37da6","Type":"ContainerDied","Data":"f28894ca2b19f1ed8167b1038b9863274e9c512b4957bb3fa094b3384519fa1c"} Oct 01 10:45:04 crc kubenswrapper[4787]: I1001 10:45:04.431377 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f28894ca2b19f1ed8167b1038b9863274e9c512b4957bb3fa094b3384519fa1c" Oct 01 10:45:04 crc kubenswrapper[4787]: I1001 10:45:04.431094 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29321925-nm6wd" Oct 01 10:45:04 crc kubenswrapper[4787]: I1001 10:45:04.479522 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321880-mxzcx"] Oct 01 10:45:04 crc kubenswrapper[4787]: I1001 10:45:04.493165 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29321880-mxzcx"] Oct 01 10:45:04 crc kubenswrapper[4787]: I1001 10:45:04.535213 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21868214-109b-4037-bacc-d984deef93fb" path="/var/lib/kubelet/pods/21868214-109b-4037-bacc-d984deef93fb/volumes" Oct 01 10:45:04 crc kubenswrapper[4787]: E1001 10:45:04.590000 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5feed863_86f3_46eb_811c_f57daec37da6.slice/crio-f28894ca2b19f1ed8167b1038b9863274e9c512b4957bb3fa094b3384519fa1c\": RecentStats: unable to find data in memory cache]" Oct 01 10:45:33 crc kubenswrapper[4787]: I1001 10:45:33.829674 4787 scope.go:117] "RemoveContainer" containerID="0bac0d9bdfd0708a883535626a9669c8a625140b488fef50fa60b3c1948cccbf" Oct 01 10:45:33 crc kubenswrapper[4787]: I1001 10:45:33.854005 4787 scope.go:117] "RemoveContainer" containerID="3c0a8d063a825052343d120852a4586d34169cfae7fa50df892b915a15b118ec" Oct 01 10:45:41 crc kubenswrapper[4787]: I1001 10:45:41.250624 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:45:41 crc kubenswrapper[4787]: I1001 10:45:41.251200 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 01 10:46:11 crc kubenswrapper[4787]: I1001 10:46:11.250537 4787 patch_prober.go:28] interesting pod/machine-config-daemon-8vh6d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 01 10:46:11 crc kubenswrapper[4787]: I1001 10:46:11.251209 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8vh6d" podUID="fa51b883-5f3e-4141-9d4c-37704aac7718" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515067203173024451 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015067203173017366 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015067172447016522 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015067172450015464 5ustar corecore